mirror of
				https://sourceware.org/git/glibc.git
				synced 2025-11-03 20:53:13 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			371 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			371 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/* Relocate a shared object and resolve its references to other loaded objects.
 | 
						|
   Copyright (C) 1995-2014 Free Software Foundation, Inc.
 | 
						|
   This file is part of the GNU C Library.
 | 
						|
 | 
						|
   The GNU C Library is free software; you can redistribute it and/or
 | 
						|
   modify it under the terms of the GNU Lesser General Public
 | 
						|
   License as published by the Free Software Foundation; either
 | 
						|
   version 2.1 of the License, or (at your option) any later version.
 | 
						|
 | 
						|
   The GNU C Library is distributed in the hope that it will be useful,
 | 
						|
   but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | 
						|
   Lesser General Public License for more details.
 | 
						|
 | 
						|
   You should have received a copy of the GNU Lesser General Public
 | 
						|
   License along with the GNU C Library; if not, see
 | 
						|
   <http://www.gnu.org/licenses/>.  */
 | 
						|
 | 
						|
#include <errno.h>
 | 
						|
#include <libintl.h>
 | 
						|
#include <stdlib.h>
 | 
						|
#include <unistd.h>
 | 
						|
#include <ldsodefs.h>
 | 
						|
#include <sys/mman.h>
 | 
						|
#include <sys/param.h>
 | 
						|
#include <sys/types.h>
 | 
						|
#include <_itoa.h>
 | 
						|
#include "dynamic-link.h"
 | 
						|
 | 
						|
/* Statistics function.  */
 | 
						|
#ifdef SHARED
 | 
						|
# define bump_num_cache_relocations() ++GL(dl_num_cache_relocations)
 | 
						|
#else
 | 
						|
# define bump_num_cache_relocations() ((void) 0)
 | 
						|
#endif
 | 
						|
 | 
						|
 | 
						|
/* We are trying to perform a static TLS relocation in MAP, but it was
 | 
						|
   dynamically loaded.  This can only work if there is enough surplus in
 | 
						|
   the static TLS area already allocated for each running thread.  If this
 | 
						|
   object's TLS segment is too big to fit, we fail.  If it fits,
 | 
						|
   we set MAP->l_tls_offset and return.
 | 
						|
   This function intentionally does not return any value but signals error
 | 
						|
   directly, as static TLS should be rare and code handling it should
 | 
						|
   not be inlined as much as possible.  */
 | 
						|
int
 | 
						|
internal_function
 | 
						|
_dl_try_allocate_static_tls (struct link_map *map)
 | 
						|
{
 | 
						|
  /* If we've already used the variable with dynamic access, or if the
 | 
						|
     alignment requirements are too high, fail.  */
 | 
						|
  if (map->l_tls_offset == FORCED_DYNAMIC_TLS_OFFSET
 | 
						|
      || map->l_tls_align > GL(dl_tls_static_align))
 | 
						|
    {
 | 
						|
    fail:
 | 
						|
      return -1;
 | 
						|
    }
 | 
						|
 | 
						|
#if TLS_TCB_AT_TP
 | 
						|
  size_t freebytes = GL(dl_tls_static_size) - GL(dl_tls_static_used);
 | 
						|
  if (freebytes < TLS_TCB_SIZE)
 | 
						|
    goto fail;
 | 
						|
  freebytes -= TLS_TCB_SIZE;
 | 
						|
 | 
						|
  size_t blsize = map->l_tls_blocksize + map->l_tls_firstbyte_offset;
 | 
						|
  if (freebytes < blsize)
 | 
						|
    goto fail;
 | 
						|
 | 
						|
  size_t n = (freebytes - blsize) / map->l_tls_align;
 | 
						|
 | 
						|
  size_t offset = GL(dl_tls_static_used) + (freebytes - n * map->l_tls_align
 | 
						|
					    - map->l_tls_firstbyte_offset);
 | 
						|
 | 
						|
  map->l_tls_offset = GL(dl_tls_static_used) = offset;
 | 
						|
#elif TLS_DTV_AT_TP
 | 
						|
  /* dl_tls_static_used includes the TCB at the beginning.  */
 | 
						|
  size_t offset = (((GL(dl_tls_static_used)
 | 
						|
		     - map->l_tls_firstbyte_offset
 | 
						|
		     + map->l_tls_align - 1) & -map->l_tls_align)
 | 
						|
		   + map->l_tls_firstbyte_offset);
 | 
						|
  size_t used = offset + map->l_tls_blocksize;
 | 
						|
 | 
						|
  if (used > GL(dl_tls_static_size))
 | 
						|
    goto fail;
 | 
						|
 | 
						|
  map->l_tls_offset = offset;
 | 
						|
  map->l_tls_firstbyte_offset = GL(dl_tls_static_used);
 | 
						|
  GL(dl_tls_static_used) = used;
 | 
						|
#else
 | 
						|
# error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined"
 | 
						|
#endif
 | 
						|
 | 
						|
  /* If the object is not yet relocated we cannot initialize the
 | 
						|
     static TLS region.  Delay it.  */
 | 
						|
  if (map->l_real->l_relocated)
 | 
						|
    {
 | 
						|
#ifdef SHARED
 | 
						|
      if (__builtin_expect (THREAD_DTV()[0].counter != GL(dl_tls_generation),
 | 
						|
			    0))
 | 
						|
	/* Update the slot information data for at least the generation of
 | 
						|
	   the DSO we are allocating data for.  */
 | 
						|
	(void) _dl_update_slotinfo (map->l_tls_modid);
 | 
						|
#endif
 | 
						|
 | 
						|
      GL(dl_init_static_tls) (map);
 | 
						|
    }
 | 
						|
  else
 | 
						|
    map->l_need_tls_init = 1;
 | 
						|
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
internal_function __attribute_noinline__
 | 
						|
_dl_allocate_static_tls (struct link_map *map)
 | 
						|
{
 | 
						|
  if (map->l_tls_offset == FORCED_DYNAMIC_TLS_OFFSET
 | 
						|
      || _dl_try_allocate_static_tls (map))
 | 
						|
    {
 | 
						|
      _dl_signal_error (0, map->l_name, NULL, N_("\
 | 
						|
cannot allocate memory in static TLS block"));
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
/* Initialize static TLS area and DTV for current (only) thread.
 | 
						|
   libpthread implementations should provide their own hook
 | 
						|
   to handle all threads.  */
 | 
						|
void
 | 
						|
_dl_nothread_init_static_tls (struct link_map *map)
 | 
						|
{
 | 
						|
#if TLS_TCB_AT_TP
 | 
						|
  void *dest = (char *) THREAD_SELF - map->l_tls_offset;
 | 
						|
#elif TLS_DTV_AT_TP
 | 
						|
  void *dest = (char *) THREAD_SELF + map->l_tls_offset + TLS_PRE_TCB_SIZE;
 | 
						|
#else
 | 
						|
# error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined"
 | 
						|
#endif
 | 
						|
 | 
						|
  /* Fill in the DTV slot so that a later LD/GD access will find it.  */
 | 
						|
  dtv_t *dtv = THREAD_DTV ();
 | 
						|
  assert (map->l_tls_modid <= dtv[-1].counter);
 | 
						|
  dtv[map->l_tls_modid].pointer.val = dest;
 | 
						|
  dtv[map->l_tls_modid].pointer.is_static = true;
 | 
						|
 | 
						|
  /* Initialize the memory.  */
 | 
						|
  memset (__mempcpy (dest, map->l_tls_initimage, map->l_tls_initimage_size),
 | 
						|
	  '\0', map->l_tls_blocksize - map->l_tls_initimage_size);
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
void
 | 
						|
_dl_relocate_object (struct link_map *l, struct r_scope_elem *scope[],
 | 
						|
		     int reloc_mode, int consider_profiling)
 | 
						|
{
 | 
						|
  struct textrels
 | 
						|
  {
 | 
						|
    caddr_t start;
 | 
						|
    size_t len;
 | 
						|
    int prot;
 | 
						|
    struct textrels *next;
 | 
						|
  } *textrels = NULL;
 | 
						|
  /* Initialize it to make the compiler happy.  */
 | 
						|
  const char *errstring = NULL;
 | 
						|
  int lazy = reloc_mode & RTLD_LAZY;
 | 
						|
  int skip_ifunc = reloc_mode & __RTLD_NOIFUNC;
 | 
						|
 | 
						|
#ifdef SHARED
 | 
						|
  /* If we are auditing, install the same handlers we need for profiling.  */
 | 
						|
  if ((reloc_mode & __RTLD_AUDIT) == 0)
 | 
						|
    consider_profiling |= GLRO(dl_audit) != NULL;
 | 
						|
#elif defined PROF
 | 
						|
  /* Never use dynamic linker profiling for gprof profiling code.  */
 | 
						|
# define consider_profiling 0
 | 
						|
#endif
 | 
						|
 | 
						|
  if (l->l_relocated)
 | 
						|
    return;
 | 
						|
 | 
						|
  /* If DT_BIND_NOW is set relocate all references in this object.  We
 | 
						|
     do not do this if we are profiling, of course.  */
 | 
						|
  // XXX Correct for auditing?
 | 
						|
  if (!consider_profiling
 | 
						|
      && __builtin_expect (l->l_info[DT_BIND_NOW] != NULL, 0))
 | 
						|
    lazy = 0;
 | 
						|
 | 
						|
  if (__glibc_unlikely (GLRO(dl_debug_mask) & DL_DEBUG_RELOC))
 | 
						|
    _dl_debug_printf ("\nrelocation processing: %s%s\n",
 | 
						|
		      DSO_FILENAME (l->l_name), lazy ? " (lazy)" : "");
 | 
						|
 | 
						|
  /* DT_TEXTREL is now in level 2 and might phase out at some time.
 | 
						|
     But we rewrite the DT_FLAGS entry to a DT_TEXTREL entry to make
 | 
						|
     testing easier and therefore it will be available at all time.  */
 | 
						|
  if (__glibc_unlikely (l->l_info[DT_TEXTREL] != NULL))
 | 
						|
    {
 | 
						|
      /* Bletch.  We must make read-only segments writable
 | 
						|
	 long enough to relocate them.  */
 | 
						|
      const ElfW(Phdr) *ph;
 | 
						|
      for (ph = l->l_phdr; ph < &l->l_phdr[l->l_phnum]; ++ph)
 | 
						|
	if (ph->p_type == PT_LOAD && (ph->p_flags & PF_W) == 0)
 | 
						|
	  {
 | 
						|
	    struct textrels *newp;
 | 
						|
 | 
						|
	    newp = (struct textrels *) alloca (sizeof (*newp));
 | 
						|
	    newp->len = (((ph->p_vaddr + ph->p_memsz + GLRO(dl_pagesize) - 1)
 | 
						|
			  & ~(GLRO(dl_pagesize) - 1))
 | 
						|
			 - (ph->p_vaddr & ~(GLRO(dl_pagesize) - 1)));
 | 
						|
	    newp->start = ((ph->p_vaddr & ~(GLRO(dl_pagesize) - 1))
 | 
						|
			   + (caddr_t) l->l_addr);
 | 
						|
 | 
						|
	    if (__mprotect (newp->start, newp->len, PROT_READ|PROT_WRITE) < 0)
 | 
						|
	      {
 | 
						|
		errstring = N_("cannot make segment writable for relocation");
 | 
						|
	      call_error:
 | 
						|
		_dl_signal_error (errno, l->l_name, NULL, errstring);
 | 
						|
	      }
 | 
						|
 | 
						|
#if (PF_R | PF_W | PF_X) == 7 && (PROT_READ | PROT_WRITE | PROT_EXEC) == 7
 | 
						|
	    newp->prot = (PF_TO_PROT
 | 
						|
			  >> ((ph->p_flags & (PF_R | PF_W | PF_X)) * 4)) & 0xf;
 | 
						|
#else
 | 
						|
	    newp->prot = 0;
 | 
						|
	    if (ph->p_flags & PF_R)
 | 
						|
	      newp->prot |= PROT_READ;
 | 
						|
	    if (ph->p_flags & PF_W)
 | 
						|
	      newp->prot |= PROT_WRITE;
 | 
						|
	    if (ph->p_flags & PF_X)
 | 
						|
	      newp->prot |= PROT_EXEC;
 | 
						|
#endif
 | 
						|
	    newp->next = textrels;
 | 
						|
	    textrels = newp;
 | 
						|
	  }
 | 
						|
    }
 | 
						|
 | 
						|
  {
 | 
						|
    /* Do the actual relocation of the object's GOT and other data.  */
 | 
						|
 | 
						|
    /* String table object symbols.  */
 | 
						|
    const char *strtab = (const void *) D_PTR (l, l_info[DT_STRTAB]);
 | 
						|
 | 
						|
    /* This macro is used as a callback from the ELF_DYNAMIC_RELOCATE code.  */
 | 
						|
#define RESOLVE_MAP(ref, version, r_type) \
 | 
						|
    (ELFW(ST_BIND) ((*ref)->st_info) != STB_LOCAL			      \
 | 
						|
     ? ((__builtin_expect ((*ref) == l->l_lookup_cache.sym, 0)		      \
 | 
						|
	 && elf_machine_type_class (r_type) == l->l_lookup_cache.type_class)  \
 | 
						|
	? (bump_num_cache_relocations (),				      \
 | 
						|
	   (*ref) = l->l_lookup_cache.ret,				      \
 | 
						|
	   l->l_lookup_cache.value)					      \
 | 
						|
	: ({ lookup_t _lr;						      \
 | 
						|
	     int _tc = elf_machine_type_class (r_type);			      \
 | 
						|
	     l->l_lookup_cache.type_class = _tc;			      \
 | 
						|
	     l->l_lookup_cache.sym = (*ref);				      \
 | 
						|
	     const struct r_found_version *v = NULL;			      \
 | 
						|
	     if ((version) != NULL && (version)->hash != 0)		      \
 | 
						|
	       v = (version);						      \
 | 
						|
	     _lr = _dl_lookup_symbol_x (strtab + (*ref)->st_name, l, (ref),   \
 | 
						|
					scope, v, _tc,			      \
 | 
						|
					DL_LOOKUP_ADD_DEPENDENCY, NULL);      \
 | 
						|
	     l->l_lookup_cache.ret = (*ref);				      \
 | 
						|
	     l->l_lookup_cache.value = _lr; }))				      \
 | 
						|
     : l)
 | 
						|
 | 
						|
#include "dynamic-link.h"
 | 
						|
 | 
						|
    ELF_DYNAMIC_RELOCATE (l, lazy, consider_profiling, skip_ifunc);
 | 
						|
 | 
						|
#ifndef PROF
 | 
						|
    if (__glibc_unlikely (consider_profiling))
 | 
						|
      {
 | 
						|
	/* Allocate the array which will contain the already found
 | 
						|
	   relocations.  If the shared object lacks a PLT (for example
 | 
						|
	   if it only contains lead function) the l_info[DT_PLTRELSZ]
 | 
						|
	   will be NULL.  */
 | 
						|
	if (l->l_info[DT_PLTRELSZ] == NULL)
 | 
						|
	  {
 | 
						|
	    errstring = N_("%s: no PLTREL found in object %s\n");
 | 
						|
	  fatal:
 | 
						|
	    _dl_fatal_printf (errstring,
 | 
						|
			      RTLD_PROGNAME,
 | 
						|
			      l->l_name);
 | 
						|
	  }
 | 
						|
 | 
						|
	l->l_reloc_result = calloc (sizeof (l->l_reloc_result[0]),
 | 
						|
				    l->l_info[DT_PLTRELSZ]->d_un.d_val);
 | 
						|
	if (l->l_reloc_result == NULL)
 | 
						|
	  {
 | 
						|
	    errstring = N_("\
 | 
						|
%s: out of memory to store relocation results for %s\n");
 | 
						|
	    goto fatal;
 | 
						|
	  }
 | 
						|
      }
 | 
						|
#endif
 | 
						|
  }
 | 
						|
 | 
						|
  /* Mark the object so we know this work has been done.  */
 | 
						|
  l->l_relocated = 1;
 | 
						|
 | 
						|
  /* Undo the segment protection changes.  */
 | 
						|
  while (__builtin_expect (textrels != NULL, 0))
 | 
						|
    {
 | 
						|
      if (__mprotect (textrels->start, textrels->len, textrels->prot) < 0)
 | 
						|
	{
 | 
						|
	  errstring = N_("cannot restore segment prot after reloc");
 | 
						|
	  goto call_error;
 | 
						|
	}
 | 
						|
 | 
						|
#ifdef CLEAR_CACHE
 | 
						|
      CLEAR_CACHE (textrels->start, textrels->start + textrels->len);
 | 
						|
#endif
 | 
						|
 | 
						|
      textrels = textrels->next;
 | 
						|
    }
 | 
						|
 | 
						|
  /* In case we can protect the data now that the relocations are
 | 
						|
     done, do it.  */
 | 
						|
  if (l->l_relro_size != 0)
 | 
						|
    _dl_protect_relro (l);
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
void internal_function
 | 
						|
_dl_protect_relro (struct link_map *l)
 | 
						|
{
 | 
						|
  ElfW(Addr) start = ((l->l_addr + l->l_relro_addr)
 | 
						|
		      & ~(GLRO(dl_pagesize) - 1));
 | 
						|
  ElfW(Addr) end = ((l->l_addr + l->l_relro_addr + l->l_relro_size)
 | 
						|
		    & ~(GLRO(dl_pagesize) - 1));
 | 
						|
 | 
						|
  if (start != end
 | 
						|
      && __mprotect ((void *) start, end - start, PROT_READ) < 0)
 | 
						|
    {
 | 
						|
      static const char errstring[] = N_("\
 | 
						|
cannot apply additional memory protection after relocation");
 | 
						|
      _dl_signal_error (errno, l->l_name, NULL, errstring);
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
internal_function __attribute_noinline__
 | 
						|
_dl_reloc_bad_type (struct link_map *map, unsigned int type, int plt)
 | 
						|
{
 | 
						|
#define DIGIT(b)	_itoa_lower_digits[(b) & 0xf];
 | 
						|
 | 
						|
  /* XXX We cannot translate these messages.  */
 | 
						|
  static const char msg[2][32
 | 
						|
#if __ELF_NATIVE_CLASS == 64
 | 
						|
			   + 6
 | 
						|
#endif
 | 
						|
  ] = { "unexpected reloc type 0x",
 | 
						|
	"unexpected PLT reloc type 0x" };
 | 
						|
  char msgbuf[sizeof (msg[0])];
 | 
						|
  char *cp;
 | 
						|
 | 
						|
  cp = __stpcpy (msgbuf, msg[plt]);
 | 
						|
#if __ELF_NATIVE_CLASS == 64
 | 
						|
  if (__builtin_expect(type > 0xff, 0))
 | 
						|
    {
 | 
						|
      *cp++ = DIGIT (type >> 28);
 | 
						|
      *cp++ = DIGIT (type >> 24);
 | 
						|
      *cp++ = DIGIT (type >> 20);
 | 
						|
      *cp++ = DIGIT (type >> 16);
 | 
						|
      *cp++ = DIGIT (type >> 12);
 | 
						|
      *cp++ = DIGIT (type >> 8);
 | 
						|
    }
 | 
						|
#endif
 | 
						|
  *cp++ = DIGIT (type >> 4);
 | 
						|
  *cp++ = DIGIT (type);
 | 
						|
  *cp = '\0';
 | 
						|
 | 
						|
  _dl_signal_error (0, map->l_name, NULL, msgbuf);
 | 
						|
}
 |