diff options
author | Ulrich Drepper <drepper@redhat.com> | 2000-05-05 07:15:29 +0000 |
---|---|---|
committer | Ulrich Drepper <drepper@redhat.com> | 2000-05-05 07:15:29 +0000 |
commit | c0282c0642e99b375ab14fd343aa537445cd72a4 (patch) | |
tree | 2fb1716922b899555ab3bdcfbd95225d7e1e912d /sysdeps/ia64/dl-machine.h | |
parent | 61d357cae37d7d84afa50bdba12dcf177980e7bf (diff) | |
download | glibc-c0282c0642e99b375ab14fd343aa537445cd72a4.tar.gz glibc-c0282c0642e99b375ab14fd343aa537445cd72a4.tar.xz glibc-c0282c0642e99b375ab14fd343aa537445cd72a4.zip |
Update.
2000-05-05 Ulrich Drepper <drepper@redhat.com> * elf/dl-load.c (_dl_map_object_from_fd): Little of computation of parameter to mprotect and for variable assignments. 2000-05-03 Jes Sorensen <jes@linuxcare.com> * sysdeps/generic/ldsodefs.h (LOOKUP_VALUE_ADDRESS): Check the validity of map before dereferencing it. * elf/dl-reloc.c (RESOLVE_MAP): Define. 2000-05-02 Jes Sorensen <jes@linuxcare.com> * elf/dl-runtime.c (fixup): Add the value returned in the symbol lookup to the arguments to elf_machine_fixup_plt(). * sysdeps/ia64/dl-machine.h (elf_machine_fixup_plt): Add Link_map of the symbol being resolved to input argument list and make the function return the pointer to the reloc. * sysdeps/alpha/dl-machine.h (elf_machine_fixup_plt): Change return valuie to lookup_t and return the value. * sysdeps/arm/dl-machine.h (elf_machine_fixup_plt): Likewise. * sysdeps/generic/dl-machine.h (elf_machine_fixup_plt): Likewise. * sysdeps/i386/dl-machine.h (elf_machine_fixup_plt): Likewise. * sysdeps/m68k/dl-machine.h (elf_machine_fixup_plt): Likewise. * sysdeps/sparc/sparc32/dl-machine.h (elf_machine_fixup_plt): Likewise. * sysdeps/sparc/sparc64/dl-machine.h (elf_machine_fixup_plt): Likewise. * sysdeps/powerpc/dl-machine.h (elf_machine_fixup_plt): Likewise. Make it an inline function returning value after calling __elf_machine_fixup_plt(). * elf/dl-sym.c (_dl_vsym): Use DL_SYMBOL_ADDRESS() to obtain the symbol address. * elf/dl-symbol.c (_dl_symbol_value): Use LOOKUP_VALUE_ADDRESS to obtain the symbol address. * sysdeps/generic/ldsodefs.h: Add generic DL_SYMBOL_ADDRESS() macro depending on the definition of ELF_FUNCTION_PTR_IS_SPECIAL. * sysdeps/ia64/dl-machine.h: Add DL_SYMBOL_ADDRESS() macro calling _dl_symbol_address() - this way DL_SYMBOL_ADDRESS() turns into an inline on non ia64. 2000-04-28 Jes Sorensen <jes@linuxcare.com> * elf/dl-runtime.c (fixup): Use the portable macros to get the symbol address of an object. * elf/dl-runtime.c (fixup-profile): Use the portable macros to get the symbol address of an object. * elf/dl-libc.c (struct do_dlsym_args): Change loadbase to a lookup_t. * elf/dl-lookup.c (_dl_lookup_symbol): Likewise. (_dl_lookup_symbol_skip): Likewise. (_dl_lookup_versioned_symbol): Likewise. (_dl_lookup_versioned_symbol_skip): Likewise. 2000-04-27 Jes Sorensen <jes@linuxcare.com> * elf/rtld.c (_dl_start): Get the function pointer return address via _dl_start_address for architectures that need a function pointer descriptor rather than just a pointer (ia64). * sysdeps/generic/dl-lookupcfg.h: New file. * sysdeps/ia64/dl-lookupcfg.h: New file. * sysdeps/ia64/dl-machine.h: New file. * sysdeps/ia64/dl-symaddr.c: New file. * sysdeps/ia64/dl-fptr.c: New file. * elf/elf.h: Add IA-64 specific definitions.
Diffstat (limited to 'sysdeps/ia64/dl-machine.h')
-rw-r--r-- | sysdeps/ia64/dl-machine.h | 585 |
1 files changed, 585 insertions, 0 deletions
diff --git a/sysdeps/ia64/dl-machine.h b/sysdeps/ia64/dl-machine.h new file mode 100644 index 0000000000..7fba08dabe --- /dev/null +++ b/sysdeps/ia64/dl-machine.h @@ -0,0 +1,585 @@ +/* Machine-dependent ELF dynamic relocation inline functions. IA-64 version. + Copyright (C) 1995, 1996, 1997, 2000 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Library General Public License as + published by the Free Software Foundation; either version 2 of the + License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Library General Public License for more details. + + You should have received a copy of the GNU Library General Public + License along with the GNU C Library; see the file COPYING.LIB. If + not, write to the Free Software Foundation, Inc., + 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ + +#ifndef dl_machine_h +#define dl_machine_h 1 + +#define ELF_MACHINE_NAME "ia64" + +#include <assert.h> +#include <string.h> +#include <link.h> +#include <errno.h> + + +/* Translate a processor specific dynamic tag to the index + in l_info array. */ +#define DT_IA_64(x) (DT_IA_64_##x - DT_LOPROC + DT_NUM) + + +/* An FPTR is a function descriptor. Properly they consist of just + FUNC and GP. But we want to traverse a binary tree too. */ + +#define IA64_BOOT_FPTR_SIZE 256 + +struct ia64_fptr +{ + Elf64_Addr func; + Elf64_Addr gp; + struct ia64_fptr *next; +}; + +extern struct ia64_fptr __boot_ldso_fptr[]; +extern struct ia64_fptr *__fptr_next; +extern struct ia64_fptr *__fptr_root; +extern int __fptr_count; + +extern Elf64_Addr __ia64_make_fptr (const struct link_map *, Elf64_Addr, + struct ia64_fptr **, struct ia64_fptr *); + +/* Return nonzero iff E_MACHINE is compatible with the running host. */ +static inline int +elf_machine_matches_host (Elf64_Word e_machine) +{ + return e_machine == EM_IA_64; +} + +void * _dl_symbol_address (const struct link_map *map, const ElfW(Sym) *ref); + +#define DL_SYMBOL_ADDRESS(map, ref) _dl_symbol_address(map, ref) + + +/* Return the link-time address of _DYNAMIC. */ +static inline Elf64_Addr +elf_machine_dynamic (void) +{ + Elf64_Addr *p; + + __asm__( + ".section .sdata\n" + " .type __dynamic_ltv#, @object\n" + " .size __dynamic_ltv#, 8\n" + "__dynamic_ltv:\n" + " data8 @ltv(_DYNAMIC#)\n" + ".previous\n" + " addl %0 = @gprel(__dynamic_ltv#), gp ;;" + : "=r"(p)); + + return *p; +} + + +/* Return the run-time load address of the shared object. */ +static inline Elf64_Addr +elf_machine_load_address (void) +{ + Elf64_Addr ip; + int *p; + + __asm__( + "1: mov %0 = ip\n" + ".section .sdata\n" + "2: data4 @ltv(1b)\n" + " .align 8\n" + ".previous\n" + " addl %1 = @gprel(2b), gp ;;" + : "=r"(ip), "=r"(p)); + + return ip - (Elf64_Addr)*p; +} + + +/* Set up the loaded object described by L so its unrelocated PLT + entries will jump to the on-demand fixup code in dl-runtime.c. */ + +static inline int +elf_machine_runtime_setup (struct link_map *l, int lazy, int profile) +{ + extern void _dl_runtime_resolve (void); + extern void _dl_runtime_profile (void); + + if (lazy) + { + register Elf64_Addr gp __asm__("gp"); + Elf64_Addr *reserve, doit; + + /* + * Careful with the typecast here or it will try to add l-l_addr + * pointer elements + */ + reserve = (Elf64_Addr *) + (l->l_info[DT_IA_64(PLT_RESERVE)]->d_un.d_ptr + l->l_addr); + /* Identify this shared object. */ + reserve[0] = (Elf64_Addr) l; + + /* This function will be called to perform the relocation. */ + if (!profile) + doit = (Elf64_Addr) ((struct ia64_fptr *)&_dl_runtime_resolve)->func; + else + { + if (_dl_name_match_p (_dl_profile, l)) + { + /* This is the object we are looking for. Say that we really + want profiling and the timers are started. */ + _dl_profile_map = l; + } + doit = (Elf64_Addr) ((struct ia64_fptr *)&_dl_runtime_profile)->func; + } + + reserve[1] = doit; + reserve[2] = gp; + } + + return lazy; +} + + +/* + This code is used in dl-runtime.c to call the `fixup' function + and then redirect to the address it returns. `fixup()' takes two + arguments, however fixup_profile() takes three. + + The ABI specifies that we will never see more than 8 input + registers to a function call, thus it is safe to simply allocate + those, and simpler than playing stack games. + - 12/09/99 Jes + */ +#define TRAMPOLINE_TEMPLATE(tramp_name, fixup_name) \ + extern void tramp_name (void); \ + asm ( "\ + .global " #tramp_name "# + .proc " #tramp_name "# +" #tramp_name ": + { .mmi + alloc loc0 = ar.pfs, 8, 2, 3, 0 + adds r2 = -144, r12 + adds r3 = -128, r12 + } + { .mii + adds r12 = -160, r12 + mov loc1 = b0 + mov out2 = b0 /* needed by fixup_profile */ + ;; + } + { .mmi + stf.spill [r2] = f8, 32 + stf.spill [r3] = f9, 32 + mov out0 = r16 + ;; + } + { .mmi + stf.spill [r2] = f10, 32 + stf.spill [r3] = f11, 32 + shl out1 = r15, 4 + ;; + } + { .mmi + stf.spill [r2] = f12, 32 + stf.spill [r3] = f13, 32 + shladd out1 = r15, 3, out1 + ;; + } + { .mmb + stf.spill [r2] = f14 + stf.spill [r3] = f15 + br.call.sptk.many b0 = " #fixup_name "# + } + { .mii + ld8 r9 = [ret0], 8 + adds r2 = 16, r12 + adds r3 = 32, r12 + ;; + } + { .mmi + ldf.fill f8 = [r2], 32 + ldf.fill f9 = [r3], 32 + mov b0 = loc1 + ;; + } + { .mmi + ldf.fill f10 = [r2], 32 + ldf.fill f11 = [r3], 32 + mov b6 = r9 + ;; + } + { .mmi + ldf.fill f12 = [r2], 32 + ldf.fill f13 = [r3], 32 + mov ar.pfs = loc0 + ;; + } + { .mmi + ldf.fill f14 = [r2], 32 + ldf.fill f15 = [r3], 32 + adds r12 = 160, r12 + ;; + } + /* An alloc is needed for the break system call to work. + We don't care about the old value of the pfs register. */ + { .mmb + alloc r2 = ar.pfs, 0, 0, 8, 0 + ld8 gp = [ret0] + br.sptk.many b6 + ;; + } + .endp " #tramp_name "#") + +#ifndef PROF +#define ELF_MACHINE_RUNTIME_TRAMPOLINE \ + TRAMPOLINE_TEMPLATE (_dl_runtime_resolve, fixup); \ + TRAMPOLINE_TEMPLATE (_dl_runtime_profile, profile_fixup); +#else +#define ELF_MACHINE_RUNTIME_TRAMPOLINE \ + TRAMPOLINE_TEMPLATE (_dl_runtime_resolve, fixup); \ + strong_alias (_dl_runtime_resolve, _dl_runtime_profile); +#endif + + +/* Initial entry point code for the dynamic linker. + The C function `_dl_start' is the real entry point; + its return value is the user program's entry point. */ + +#define RTLD_START asm ("\ +.text + .global _start# + .proc _start# +_start: +0: { .mii + alloc loc0 = ar.pfs, 0, 3, 4, 0 + mov r2 = ip + addl r3 = @gprel(0b), r0 + ;; + } + { .mlx + /* Calculate the GP, and save a copy in loc1. */ + sub gp = r2, r3 + movl r8 = 0x9804c0270033f + ;; + } + { .mii + mov ar.fpsr = r8 + sub loc1 = r2, r3 + /* _dl_start wants a pointer to the pointer to the arg block + and the arg block starts with an integer, thus the magic 16. */ + adds out0 = 16, sp + } + { .bbb + br.call.sptk.many b0 = _dl_start# + ;; + } + .endp _start# + /* FALLTHRU */ + .global _dl_start_user# + .proc _dl_start_user# +_dl_start_user: + { .mii + /* Save the pointer to the user entry point fptr in loc2. */ + mov loc2 = ret0 + /* Store the highest stack address. */ + addl r2 = @ltoff(__libc_stack_end#), gp + addl r3 = @gprel(_dl_skip_args), gp + ;; + } + { .mmi + ld8 r2 = [r2] + ld4 r3 = [r3] + adds r11 = 24, sp /* Load the address of argv. */ + ;; + } + { .mii + st8 [r2] = sp + adds r10 = 16, sp /* Load the address of argc. */ + mov out2 = r11 + ;; + /* See if we were run as a command with the executable file + name as an extra leading argument. If so, adjust the argv + pointer to skip _dl_skip_args words. + Note that _dl_skip_args is an integer, not a long - Jes + + The stack pointer has to be 16 byte aligned. We cannot simply + addjust the stack pointer. We have to move the whole argv and + envp. H.J. */ + } + { .mmi + ld8 out1 = [r10] /* is argc actually stored as a long + or as an int? */ + ;; + sub out1 = out1, r3 /* Get the new argc. */ + shladd r15 = r3, 3, r11 /* The address of the argv we move */ + ;; + } + /* ??? Could probably merge these two loops into 3 bundles. + using predication to control which set of copies we're on. */ +1: /* Copy argv. */ + { .mfi + ld8 r16 = [r15], 8 /* Load the value in the old argv. */ + ;; + } + { .mib + st8 [r11] = r16, 8 /* Store it in the new argv. */ + cmp.ne p6, p7 = 0, r16 +(p6) br.cond.dptk.few 1b + ;; + } + { .mib + mov out3 = r11 + addl out0 = @ltoff(_dl_loaded), gp + } +1: /* Copy env. */ + { .mfi + ld8 r16 = [r15], 8 /* Load the value in the old env. */ + ;; + } + { .mib + st8 [r11] = r16, 8 /* Store it in the new env. */ + cmp.ne p6, p7 = 0, r16 +(p6) br.cond.dptk.few 1b + ;; + } + { .mmb + st8 [r10] = out1 /* Record the new argc. */ + ld8 out0 = [out0] + } + { .mfb + ld8 out0 = [out0] /* get the linkmap */ + br.call.sptk.many b0 = _dl_init# + ;; + } + /* Pass our finializer function to the user, + and jump to the user's entry point. */ + { .mmi + ld8 r3 = [loc2], 8 + mov b0 = r0 + } + { .mmi + addl ret0 = @ltoff(@fptr(_dl_fini#)), gp + ;; + mov b6 = r3 + } + { .mmi + ld8 ret0 = [ret0] + ld8 gp = [loc2] + mov ar.pfs = loc0 + ;; + } + { .mfb + br.sptk.many b6 + ;; + } + .endp _dl_start_user# +.previous"); + + +#ifndef RTLD_START_SPECIAL_INIT +#define RTLD_START_SPECIAL_INIT /* nothing */ +#endif + +/* Nonzero iff TYPE describes relocation of a PLT entry, so + PLT entries should not be allowed to define the value. */ +/* ??? Ignore IPLTMSB for now. */ +#define elf_machine_lookup_noplt_p(type) ((type) == R_IA64_IPLTLSB) + +/* Nonzero iff TYPE should not be allowed to resolve to one of + the main executable's symbols, as for a COPY reloc, which we don't use. */ +#define elf_machine_lookup_noexec_p(type) (0) + +/* A reloc type used for ld.so cmdline arg lookups to reject PLT entries. */ +#define ELF_MACHINE_JMP_SLOT R_IA64_IPLTLSB + +/* According to the IA-64 specific documentation, Rela is always used. */ +#define ELF_MACHINE_NO_REL 1 + +/* Since ia64's stack has to be 16byte aligned, we cannot arbitrarily + move the stack pointer. */ +#define ELF_MACHINE_FIXED_STACK 1 + +/* Return the address of the entry point. */ +extern ElfW(Addr) _dl_start_address (const struct link_map *map, + ElfW(Addr) start); + +#define ELF_MACHINE_START_ADDRESS(map, start) \ + _dl_start_address ((map), (start)) + +#define elf_machine_profile_fixup_plt(l, reloc, rel_addr, value) \ + elf_machine_fixup_plt ((l), (reloc), (rel_addr), (value)) + +#define elf_machine_profile_plt(reloc_addr) ((Elf64_Addr) (reloc_addr)) + +/* Fixup a PLT entry to bounce directly to the function at VALUE. */ +static inline Elf64_Addr +elf_machine_fixup_plt (struct link_map *l, lookup_t t, + const Elf64_Rela *reloc, + Elf64_Addr *reloc_addr, Elf64_Addr value) +{ + /* l is the link_map for the caller, t is the link_map for the object + * being called */ + /* got has already been relocated in elf_get_dynamic_info() */ + reloc_addr[1] = t->l_info[DT_PLTGOT]->d_un.d_ptr; + reloc_addr[0] = value; + return (Elf64_Addr) reloc_addr; +} + +/* Return the final value of a plt relocation. */ +static inline Elf64_Addr +elf_machine_plt_value (struct link_map *map, const Elf64_Rela *reloc, + Elf64_Addr value) +{ + /* No need to handle rel vs rela since IA64 is rela only */ + return value + reloc->r_addend; +} + +#endif /* !dl_machine_h */ + +#ifdef RESOLVE_MAP + +#define R_IA64_TYPE(R) ((R) & -8) +#define R_IA64_FORMAT(R) ((R) & 7) + +#define R_IA64_FORMAT_32MSB 4 +#define R_IA64_FORMAT_32LSB 5 +#define R_IA64_FORMAT_64MSB 6 +#define R_IA64_FORMAT_64LSB 7 + + +/* Perform the relocation specified by RELOC and SYM (which is fully + resolved). MAP is the object containing the reloc. */ +static inline void +elf_machine_rela (struct link_map *map, + const Elf64_Rela *reloc, + const Elf64_Sym *sym, + const struct r_found_version *version, + Elf64_Addr *const reloc_addr) +{ + unsigned long const r_type = ELF64_R_TYPE (reloc->r_info); + Elf64_Addr value; + +#ifndef RTLD_BOOTSTRAP + /* This is defined in rtld.c, but nowhere in the static libc.a; make the + reference weak so static programs can still link. This declaration + cannot be done when compiling rtld.c (i.e. #ifdef RTLD_BOOTSTRAP) + because rtld.c contains the common defn for _dl_rtld_map, which is + incompatible with a weak decl in the same file. */ + weak_extern (_dl_rtld_map); +#endif + + /* We cannot use a switch here because we cannot locate the switch + jump table until we've self-relocated. */ + + if (R_IA64_TYPE (r_type) == R_IA64_TYPE (R_IA64_REL64LSB)) + { + value = *reloc_addr; +#ifndef RTLD_BOOTSTRAP + /* Already done in dynamic linker. */ + if (map != &_dl_rtld_map) +#endif + value += map->l_addr; + } + else if (r_type == R_IA64_NONE) + return; + else + { + struct link_map *sym_map; + + /* + * RESOLVE_MAP() will return NULL if it fail to locate the symbol + */ + if ((sym_map = RESOLVE_MAP (&sym, version, r_type))) + { + value = sym ? sym_map->l_addr + sym->st_value : 0; + value += reloc->r_addend; + + if (R_IA64_TYPE (r_type) == R_IA64_TYPE (R_IA64_DIR64LSB)) + ;/* No adjustment. */ + else if (r_type == R_IA64_IPLTLSB) + { + elf_machine_fixup_plt (NULL, sym_map, reloc, reloc_addr, value); + return; + } + else if (R_IA64_TYPE (r_type) == R_IA64_TYPE (R_IA64_FPTR64LSB)) +#ifndef RTLD_BOOTSTRAP + value = __ia64_make_fptr (sym_map, value, &__fptr_root, NULL); +#else + { + struct ia64_fptr *p_boot_ldso_fptr; + struct ia64_fptr **p_fptr_root; + int *p_fptr_count; + + /* Special care must be taken to address these variables + during bootstrap. Further, since we don't know exactly + when __fptr_next will be relocated, we index directly + off __boot_ldso_fptr. */ + asm ("addl %0 = @gprel(__boot_ldso_fptr#), gp\n\t" + "addl %1 = @gprel(__fptr_root#), gp\n\t" + "addl %2 = @gprel(__fptr_count#), gp" + : "=r"(p_boot_ldso_fptr), + "=r"(p_fptr_root), + "=r"(p_fptr_count)); + + /* + * Go from the top - __ia64_make_fptr goes from the bottom, + * this way we will never clash. + */ + value = __ia64_make_fptr (sym_map, value, p_fptr_root, + &p_boot_ldso_fptr[--*p_fptr_count]); + } +#endif + else if (R_IA64_TYPE (r_type) == R_IA64_TYPE (R_IA64_PCREL64LSB)) + value -= (Elf64_Addr)reloc_addr & -16; + else + assert (! "unexpected dynamic reloc type"); + } + else + value = 0; + } + + /* ??? Ignore MSB and Instruction format for now. */ + if (R_IA64_FORMAT (r_type) == R_IA64_FORMAT_64LSB) + *reloc_addr = value; + else if (R_IA64_FORMAT (r_type) == R_IA64_FORMAT_32LSB) + *(int *)reloc_addr = value; + else if (r_type == R_IA64_IPLTLSB) + { + reloc_addr[0] = 0; + reloc_addr[1] = 0; + } + else + assert (! "unexpected dynamic reloc format"); +} + + +/* Perform a RELATIVE reloc on the .got entry that transfers to the .plt. */ +static inline void +elf_machine_lazy_rel (struct link_map *map, + Elf64_Addr l_addr, const Elf64_Rela *reloc) +{ + Elf64_Addr * const reloc_addr = (void *)(l_addr + reloc->r_offset); + unsigned long const r_type = ELF64_R_TYPE (reloc->r_info); + + if (r_type == R_IA64_IPLTLSB) + { + reloc_addr[0] += l_addr; + reloc_addr[1] += l_addr; + } + else if (r_type == R_IA64_NONE) + return; + else + assert (! "unexpected PLT reloc type"); +} + +#endif /* RESOLVE_MAP */ |