| /* Relocate a shared object and resolve its references to other loaded objects. | 
 |    Copyright (C) 1995-2016 Free Software Foundation, Inc. | 
 |    This file is part of the GNU C Library. | 
 |  | 
 |    The GNU C Library is free software; you can redistribute it and/or | 
 |    modify it under the terms of the GNU Lesser General Public | 
 |    License as published by the Free Software Foundation; either | 
 |    version 2.1 of the License, or (at your option) any later version. | 
 |  | 
 |    The GNU C Library is distributed in the hope that it will be useful, | 
 |    but WITHOUT ANY WARRANTY; without even the implied warranty of | 
 |    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU | 
 |    Lesser General Public License for more details. | 
 |  | 
 |    You should have received a copy of the GNU Lesser General Public | 
 |    License along with the GNU C Library; if not, see | 
 |    <http://www.gnu.org/licenses/>.  */ | 
 |  | 
 | #include <errno.h> | 
 | #include <libintl.h> | 
 | #include <stdlib.h> | 
 | #include <unistd.h> | 
 | #include <ldsodefs.h> | 
 | #include <sys/mman.h> | 
 | #include <sys/param.h> | 
 | #include <sys/types.h> | 
 | #include <_itoa.h> | 
 | #include "dynamic-link.h" | 
 | #include <libc-internal.h> | 
 |  | 
 | /* Statistics function.  */ | 
 | #ifdef SHARED | 
 | # define bump_num_cache_relocations() ++GL(dl_num_cache_relocations) | 
 | #else | 
 | # define bump_num_cache_relocations() ((void) 0) | 
 | #endif | 
 |  | 
 |  | 
 | /* We are trying to perform a static TLS relocation in MAP, but it was | 
 |    dynamically loaded.  This can only work if there is enough surplus in | 
 |    the static TLS area already allocated for each running thread.  If this | 
 |    object's TLS segment is too big to fit, we fail.  If it fits, | 
 |    we set MAP->l_tls_offset and return. | 
 |    This function intentionally does not return any value but signals error | 
 |    directly, as static TLS should be rare and code handling it should | 
 |    not be inlined as much as possible.  */ | 
 | int | 
 | internal_function | 
 | _dl_try_allocate_static_tls (struct link_map *map) | 
 | { | 
 |   /* If we've already used the variable with dynamic access, or if the | 
 |      alignment requirements are too high, fail.  */ | 
 |   if (map->l_tls_offset == FORCED_DYNAMIC_TLS_OFFSET | 
 |       || map->l_tls_align > GL(dl_tls_static_align)) | 
 |     { | 
 |     fail: | 
 |       return -1; | 
 |     } | 
 |  | 
 | #if TLS_TCB_AT_TP | 
 |   size_t freebytes = GL(dl_tls_static_size) - GL(dl_tls_static_used); | 
 |   if (freebytes < TLS_TCB_SIZE) | 
 |     goto fail; | 
 |   freebytes -= TLS_TCB_SIZE; | 
 |  | 
 |   size_t blsize = map->l_tls_blocksize + map->l_tls_firstbyte_offset; | 
 |   if (freebytes < blsize) | 
 |     goto fail; | 
 |  | 
 |   size_t n = (freebytes - blsize) / map->l_tls_align; | 
 |  | 
 |   size_t offset = GL(dl_tls_static_used) + (freebytes - n * map->l_tls_align | 
 | 					    - map->l_tls_firstbyte_offset); | 
 |  | 
 |   map->l_tls_offset = GL(dl_tls_static_used) = offset; | 
 | #elif TLS_DTV_AT_TP | 
 |   /* dl_tls_static_used includes the TCB at the beginning.  */ | 
 |   size_t offset = (ALIGN_UP(GL(dl_tls_static_used) | 
 | 			    - map->l_tls_firstbyte_offset, | 
 | 			    map->l_tls_align) | 
 | 		   + map->l_tls_firstbyte_offset); | 
 |   size_t used = offset + map->l_tls_blocksize; | 
 |  | 
 |   if (used > GL(dl_tls_static_size)) | 
 |     goto fail; | 
 |  | 
 |   map->l_tls_offset = offset; | 
 |   map->l_tls_firstbyte_offset = GL(dl_tls_static_used); | 
 |   GL(dl_tls_static_used) = used; | 
 | #else | 
 | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
 | #endif | 
 |  | 
 |   /* If the object is not yet relocated we cannot initialize the | 
 |      static TLS region.  Delay it.  */ | 
 |   if (map->l_real->l_relocated) | 
 |     { | 
 | #ifdef SHARED | 
 |       if (__builtin_expect (THREAD_DTV()[0].counter != GL(dl_tls_generation), | 
 | 			    0)) | 
 | 	/* Update the slot information data for at least the generation of | 
 | 	   the DSO we are allocating data for.  */ | 
 | 	(void) _dl_update_slotinfo (map->l_tls_modid); | 
 | #endif | 
 |  | 
 |       GL(dl_init_static_tls) (map); | 
 |     } | 
 |   else | 
 |     map->l_need_tls_init = 1; | 
 |  | 
 |   return 0; | 
 | } | 
 |  | 
 | void | 
 | internal_function __attribute_noinline__ | 
 | _dl_allocate_static_tls (struct link_map *map) | 
 | { | 
 |   if (map->l_tls_offset == FORCED_DYNAMIC_TLS_OFFSET | 
 |       || _dl_try_allocate_static_tls (map)) | 
 |     { | 
 |       _dl_signal_error (0, map->l_name, NULL, N_("\ | 
 | cannot allocate memory in static TLS block")); | 
 |     } | 
 | } | 
 |  | 
 | /* Initialize static TLS area and DTV for current (only) thread. | 
 |    libpthread implementations should provide their own hook | 
 |    to handle all threads.  */ | 
 | void | 
 | _dl_nothread_init_static_tls (struct link_map *map) | 
 | { | 
 | #if TLS_TCB_AT_TP | 
 |   void *dest = (char *) THREAD_SELF - map->l_tls_offset; | 
 | #elif TLS_DTV_AT_TP | 
 |   void *dest = (char *) THREAD_SELF + map->l_tls_offset + TLS_PRE_TCB_SIZE; | 
 | #else | 
 | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
 | #endif | 
 |  | 
 |   /* Initialize the memory.  */ | 
 |   memset (__mempcpy (dest, map->l_tls_initimage, map->l_tls_initimage_size), | 
 | 	  '\0', map->l_tls_blocksize - map->l_tls_initimage_size); | 
 | } | 
 |  | 
 |  | 
 | void | 
 | _dl_relocate_object (struct link_map *l, struct r_scope_elem *scope[], | 
 | 		     int reloc_mode, int consider_profiling) | 
 | { | 
 |   struct textrels | 
 |   { | 
 |     caddr_t start; | 
 |     size_t len; | 
 |     int prot; | 
 |     struct textrels *next; | 
 |   } *textrels = NULL; | 
 |   /* Initialize it to make the compiler happy.  */ | 
 |   const char *errstring = NULL; | 
 |   int lazy = reloc_mode & RTLD_LAZY; | 
 |   int skip_ifunc = reloc_mode & __RTLD_NOIFUNC; | 
 |  | 
 | #ifdef SHARED | 
 |   /* If we are auditing, install the same handlers we need for profiling.  */ | 
 |   if ((reloc_mode & __RTLD_AUDIT) == 0) | 
 |     consider_profiling |= GLRO(dl_audit) != NULL; | 
 | #elif defined PROF | 
 |   /* Never use dynamic linker profiling for gprof profiling code.  */ | 
 | # define consider_profiling 0 | 
 | #endif | 
 |  | 
 |   if (l->l_relocated) | 
 |     return; | 
 |  | 
 |   /* If DT_BIND_NOW is set relocate all references in this object.  We | 
 |      do not do this if we are profiling, of course.  */ | 
 |   // XXX Correct for auditing? | 
 |   if (!consider_profiling | 
 |       && __builtin_expect (l->l_info[DT_BIND_NOW] != NULL, 0)) | 
 |     lazy = 0; | 
 |  | 
 |   if (__glibc_unlikely (GLRO(dl_debug_mask) & DL_DEBUG_RELOC)) | 
 |     _dl_debug_printf ("\nrelocation processing: %s%s\n", | 
 | 		      DSO_FILENAME (l->l_name), lazy ? " (lazy)" : ""); | 
 |  | 
 |   /* DT_TEXTREL is now in level 2 and might phase out at some time. | 
 |      But we rewrite the DT_FLAGS entry to a DT_TEXTREL entry to make | 
 |      testing easier and therefore it will be available at all time.  */ | 
 |   if (__glibc_unlikely (l->l_info[DT_TEXTREL] != NULL)) | 
 |     { | 
 |       /* Bletch.  We must make read-only segments writable | 
 | 	 long enough to relocate them.  */ | 
 |       const ElfW(Phdr) *ph; | 
 |       for (ph = l->l_phdr; ph < &l->l_phdr[l->l_phnum]; ++ph) | 
 | 	if (ph->p_type == PT_LOAD && (ph->p_flags & PF_W) == 0) | 
 | 	  { | 
 | 	    struct textrels *newp; | 
 |  | 
 | 	    newp = (struct textrels *) alloca (sizeof (*newp)); | 
 | 	    newp->len = ALIGN_UP (ph->p_vaddr + ph->p_memsz, GLRO(dl_pagesize)) | 
 | 			- ALIGN_DOWN (ph->p_vaddr, GLRO(dl_pagesize)); | 
 | 	    newp->start = PTR_ALIGN_DOWN (ph->p_vaddr, GLRO(dl_pagesize)) | 
 | 			  + (caddr_t) l->l_addr; | 
 |  | 
 | 	    if (__mprotect (newp->start, newp->len, PROT_READ|PROT_WRITE) < 0) | 
 | 	      { | 
 | 		errstring = N_("cannot make segment writable for relocation"); | 
 | 	      call_error: | 
 | 		_dl_signal_error (errno, l->l_name, NULL, errstring); | 
 | 	      } | 
 |  | 
 | #if (PF_R | PF_W | PF_X) == 7 && (PROT_READ | PROT_WRITE | PROT_EXEC) == 7 | 
 | 	    newp->prot = (PF_TO_PROT | 
 | 			  >> ((ph->p_flags & (PF_R | PF_W | PF_X)) * 4)) & 0xf; | 
 | #else | 
 | 	    newp->prot = 0; | 
 | 	    if (ph->p_flags & PF_R) | 
 | 	      newp->prot |= PROT_READ; | 
 | 	    if (ph->p_flags & PF_W) | 
 | 	      newp->prot |= PROT_WRITE; | 
 | 	    if (ph->p_flags & PF_X) | 
 | 	      newp->prot |= PROT_EXEC; | 
 | #endif | 
 | 	    newp->next = textrels; | 
 | 	    textrels = newp; | 
 | 	  } | 
 |     } | 
 |  | 
 |   { | 
 |     /* Do the actual relocation of the object's GOT and other data.  */ | 
 |  | 
 |     /* String table object symbols.  */ | 
 |     const char *strtab = (const void *) D_PTR (l, l_info[DT_STRTAB]); | 
 |  | 
 |     /* This macro is used as a callback from the ELF_DYNAMIC_RELOCATE code.  */ | 
 | #define RESOLVE_MAP(ref, version, r_type) \ | 
 |     (ELFW(ST_BIND) ((*ref)->st_info) != STB_LOCAL			      \ | 
 |      ? ((__builtin_expect ((*ref) == l->l_lookup_cache.sym, 0)		      \ | 
 | 	 && elf_machine_type_class (r_type) == l->l_lookup_cache.type_class)  \ | 
 | 	? (bump_num_cache_relocations (),				      \ | 
 | 	   (*ref) = l->l_lookup_cache.ret,				      \ | 
 | 	   l->l_lookup_cache.value)					      \ | 
 | 	: ({ lookup_t _lr;						      \ | 
 | 	     int _tc = elf_machine_type_class (r_type);			      \ | 
 | 	     l->l_lookup_cache.type_class = _tc;			      \ | 
 | 	     l->l_lookup_cache.sym = (*ref);				      \ | 
 | 	     const struct r_found_version *v = NULL;			      \ | 
 | 	     if ((version) != NULL && (version)->hash != 0)		      \ | 
 | 	       v = (version);						      \ | 
 | 	     _lr = _dl_lookup_symbol_x (strtab + (*ref)->st_name, l, (ref),   \ | 
 | 					scope, v, _tc,			      \ | 
 | 					DL_LOOKUP_ADD_DEPENDENCY, NULL);      \ | 
 | 	     l->l_lookup_cache.ret = (*ref);				      \ | 
 | 	     l->l_lookup_cache.value = _lr; }))				      \ | 
 |      : l) | 
 |  | 
 | #include "dynamic-link.h" | 
 |  | 
 |     ELF_DYNAMIC_RELOCATE (l, lazy, consider_profiling, skip_ifunc); | 
 |  | 
 | #ifndef PROF | 
 |     if (__glibc_unlikely (consider_profiling) | 
 | 	&& l->l_info[DT_PLTRELSZ] != NULL) | 
 |       { | 
 | 	/* Allocate the array which will contain the already found | 
 | 	   relocations.  If the shared object lacks a PLT (for example | 
 | 	   if it only contains lead function) the l_info[DT_PLTRELSZ] | 
 | 	   will be NULL.  */ | 
 | 	size_t sizeofrel = l->l_info[DT_PLTREL]->d_un.d_val == DT_RELA | 
 | 			   ? sizeof (ElfW(Rela)) | 
 | 			   : sizeof (ElfW(Rel)); | 
 | 	size_t relcount = l->l_info[DT_PLTRELSZ]->d_un.d_val / sizeofrel; | 
 | 	l->l_reloc_result = calloc (sizeof (l->l_reloc_result[0]), relcount); | 
 |  | 
 | 	if (l->l_reloc_result == NULL) | 
 | 	  { | 
 | 	    errstring = N_("\ | 
 | %s: out of memory to store relocation results for %s\n"); | 
 | 	    _dl_fatal_printf (errstring, RTLD_PROGNAME, l->l_name); | 
 | 	  } | 
 |       } | 
 | #endif | 
 |   } | 
 |  | 
 |   /* Mark the object so we know this work has been done.  */ | 
 |   l->l_relocated = 1; | 
 |  | 
 |   /* Undo the segment protection changes.  */ | 
 |   while (__builtin_expect (textrels != NULL, 0)) | 
 |     { | 
 |       if (__mprotect (textrels->start, textrels->len, textrels->prot) < 0) | 
 | 	{ | 
 | 	  errstring = N_("cannot restore segment prot after reloc"); | 
 | 	  goto call_error; | 
 | 	} | 
 |  | 
 | #ifdef CLEAR_CACHE | 
 |       CLEAR_CACHE (textrels->start, textrels->start + textrels->len); | 
 | #endif | 
 |  | 
 |       textrels = textrels->next; | 
 |     } | 
 |  | 
 |   /* In case we can protect the data now that the relocations are | 
 |      done, do it.  */ | 
 |   if (l->l_relro_size != 0) | 
 |     _dl_protect_relro (l); | 
 | } | 
 |  | 
 |  | 
 | void internal_function | 
 | _dl_protect_relro (struct link_map *l) | 
 | { | 
 |   ElfW(Addr) start = ALIGN_DOWN((l->l_addr | 
 | 				 + l->l_relro_addr), | 
 | 				GLRO(dl_pagesize)); | 
 |   ElfW(Addr) end = ALIGN_DOWN((l->l_addr | 
 | 			       + l->l_relro_addr | 
 | 			       + l->l_relro_size), | 
 | 			      GLRO(dl_pagesize)); | 
 |   if (start != end | 
 |       && __mprotect ((void *) start, end - start, PROT_READ) < 0) | 
 |     { | 
 |       static const char errstring[] = N_("\ | 
 | cannot apply additional memory protection after relocation"); | 
 |       _dl_signal_error (errno, l->l_name, NULL, errstring); | 
 |     } | 
 | } | 
 |  | 
 | void | 
 | internal_function __attribute_noinline__ | 
 | _dl_reloc_bad_type (struct link_map *map, unsigned int type, int plt) | 
 | { | 
 | #define DIGIT(b)	_itoa_lower_digits[(b) & 0xf]; | 
 |  | 
 |   /* XXX We cannot translate these messages.  */ | 
 |   static const char msg[2][32 | 
 | #if __ELF_NATIVE_CLASS == 64 | 
 | 			   + 6 | 
 | #endif | 
 |   ] = { "unexpected reloc type 0x", | 
 | 	"unexpected PLT reloc type 0x" }; | 
 |   char msgbuf[sizeof (msg[0])]; | 
 |   char *cp; | 
 |  | 
 |   cp = __stpcpy (msgbuf, msg[plt]); | 
 | #if __ELF_NATIVE_CLASS == 64 | 
 |   if (__builtin_expect(type > 0xff, 0)) | 
 |     { | 
 |       *cp++ = DIGIT (type >> 28); | 
 |       *cp++ = DIGIT (type >> 24); | 
 |       *cp++ = DIGIT (type >> 20); | 
 |       *cp++ = DIGIT (type >> 16); | 
 |       *cp++ = DIGIT (type >> 12); | 
 |       *cp++ = DIGIT (type >> 8); | 
 |     } | 
 | #endif | 
 |   *cp++ = DIGIT (type >> 4); | 
 |   *cp++ = DIGIT (type); | 
 |   *cp = '\0'; | 
 |  | 
 |   _dl_signal_error (0, map->l_name, NULL, msgbuf); | 
 | } |