| /* Thread-local storage handling in the ELF dynamic linker.  Generic version. | 
 |    Copyright (C) 2002-2016 Free Software Foundation, Inc. | 
 |    This file is part of the GNU C Library. | 
 |  | 
 |    The GNU C Library is free software; you can redistribute it and/or | 
 |    modify it under the terms of the GNU Lesser General Public | 
 |    License as published by the Free Software Foundation; either | 
 |    version 2.1 of the License, or (at your option) any later version. | 
 |  | 
 |    The GNU C Library is distributed in the hope that it will be useful, | 
 |    but WITHOUT ANY WARRANTY; without even the implied warranty of | 
 |    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU | 
 |    Lesser General Public License for more details. | 
 |  | 
 |    You should have received a copy of the GNU Lesser General Public | 
 |    License along with the GNU C Library; if not, see | 
 |    <http://www.gnu.org/licenses/>.  */ | 
 |  | 
 | #include <assert.h> | 
 | #include <errno.h> | 
 | #include <libintl.h> | 
 | #include <signal.h> | 
 | #include <stdlib.h> | 
 | #include <unistd.h> | 
 | #include <sys/param.h> | 
 | #include <atomic.h> | 
 |  | 
 | #include <tls.h> | 
 | #include <dl-tls.h> | 
 | #include <ldsodefs.h> | 
 |  | 
 | /* Amount of excess space to allocate in the static TLS area | 
 |    to allow dynamic loading of modules defining IE-model TLS data.  */ | 
 | #define TLS_STATIC_SURPLUS	64 + DL_NNS * 100 | 
 |  | 
 |  | 
 | /* Out-of-memory handler.  */ | 
 | static void | 
 | __attribute__ ((__noreturn__)) | 
 | oom (void) | 
 | { | 
 |   _dl_fatal_printf ("cannot allocate memory for thread-local data: ABORT\n"); | 
 | } | 
 |  | 
 |  | 
 | size_t | 
 | internal_function | 
 | _dl_next_tls_modid (void) | 
 | { | 
 |   size_t result; | 
 |  | 
 |   if (__builtin_expect (GL(dl_tls_dtv_gaps), false)) | 
 |     { | 
 |       size_t disp = 0; | 
 |       struct dtv_slotinfo_list *runp = GL(dl_tls_dtv_slotinfo_list); | 
 |  | 
 |       /* Note that this branch will never be executed during program | 
 | 	 start since there are no gaps at that time.  Therefore it | 
 | 	 does not matter that the dl_tls_dtv_slotinfo is not allocated | 
 | 	 yet when the function is called for the first times. | 
 |  | 
 | 	 NB: the offset +1 is due to the fact that DTV[0] is used | 
 | 	 for something else.  */ | 
 |       result = GL(dl_tls_static_nelem) + 1; | 
 |       if (result <= GL(dl_tls_max_dtv_idx)) | 
 | 	do | 
 | 	  { | 
 | 	    while (result - disp < runp->len) | 
 | 	      { | 
 | 		if (runp->slotinfo[result - disp].map == NULL) | 
 | 		  break; | 
 |  | 
 | 		++result; | 
 | 		assert (result <= GL(dl_tls_max_dtv_idx) + 1); | 
 | 	      } | 
 |  | 
 | 	    if (result - disp < runp->len) | 
 | 	      break; | 
 |  | 
 | 	    disp += runp->len; | 
 | 	  } | 
 | 	while ((runp = runp->next) != NULL); | 
 |  | 
 |       if (result > GL(dl_tls_max_dtv_idx)) | 
 | 	{ | 
 | 	  /* The new index must indeed be exactly one higher than the | 
 | 	     previous high.  */ | 
 | 	  assert (result == GL(dl_tls_max_dtv_idx) + 1); | 
 | 	  /* There is no gap anymore.  */ | 
 | 	  GL(dl_tls_dtv_gaps) = false; | 
 |  | 
 | 	  goto nogaps; | 
 | 	} | 
 |     } | 
 |   else | 
 |     { | 
 |       /* No gaps, allocate a new entry.  */ | 
 |     nogaps: | 
 |  | 
 |       result = ++GL(dl_tls_max_dtv_idx); | 
 |     } | 
 |  | 
 |   return result; | 
 | } | 
 |  | 
 |  | 
 | size_t | 
 | internal_function | 
 | _dl_count_modids (void) | 
 | { | 
 |   /* It is rare that we have gaps; see elf/dl-open.c (_dl_open) where | 
 |      we fail to load a module and unload it leaving a gap.  If we don't | 
 |      have gaps then the number of modids is the current maximum so | 
 |      return that.  */ | 
 |   if (__glibc_likely (!GL(dl_tls_dtv_gaps))) | 
 |     return GL(dl_tls_max_dtv_idx); | 
 |  | 
 |   /* We have gaps and are forced to count the non-NULL entries.  */ | 
 |   size_t n = 0; | 
 |   struct dtv_slotinfo_list *runp = GL(dl_tls_dtv_slotinfo_list); | 
 |   while (runp != NULL) | 
 |     { | 
 |       for (size_t i = 0; i < runp->len; ++i) | 
 | 	if (runp->slotinfo[i].map != NULL) | 
 | 	  ++n; | 
 |  | 
 |       runp = runp->next; | 
 |     } | 
 |  | 
 |   return n; | 
 | } | 
 |  | 
 |  | 
 | #ifdef SHARED | 
 | void | 
 | internal_function | 
 | _dl_determine_tlsoffset (void) | 
 | { | 
 |   size_t max_align = TLS_TCB_ALIGN; | 
 |   size_t freetop = 0; | 
 |   size_t freebottom = 0; | 
 |  | 
 |   /* The first element of the dtv slot info list is allocated.  */ | 
 |   assert (GL(dl_tls_dtv_slotinfo_list) != NULL); | 
 |   /* There is at this point only one element in the | 
 |      dl_tls_dtv_slotinfo_list list.  */ | 
 |   assert (GL(dl_tls_dtv_slotinfo_list)->next == NULL); | 
 |  | 
 |   struct dtv_slotinfo *slotinfo = GL(dl_tls_dtv_slotinfo_list)->slotinfo; | 
 |  | 
 |   /* Determining the offset of the various parts of the static TLS | 
 |      block has several dependencies.  In addition we have to work | 
 |      around bugs in some toolchains. | 
 |  | 
 |      Each TLS block from the objects available at link time has a size | 
 |      and an alignment requirement.  The GNU ld computes the alignment | 
 |      requirements for the data at the positions *in the file*, though. | 
 |      I.e, it is not simply possible to allocate a block with the size | 
 |      of the TLS program header entry.  The data is layed out assuming | 
 |      that the first byte of the TLS block fulfills | 
 |  | 
 |        p_vaddr mod p_align == &TLS_BLOCK mod p_align | 
 |  | 
 |      This means we have to add artificial padding at the beginning of | 
 |      the TLS block.  These bytes are never used for the TLS data in | 
 |      this module but the first byte allocated must be aligned | 
 |      according to mod p_align == 0 so that the first byte of the TLS | 
 |      block is aligned according to p_vaddr mod p_align.  This is ugly | 
 |      and the linker can help by computing the offsets in the TLS block | 
 |      assuming the first byte of the TLS block is aligned according to | 
 |      p_align. | 
 |  | 
 |      The extra space which might be allocated before the first byte of | 
 |      the TLS block need not go unused.  The code below tries to use | 
 |      that memory for the next TLS block.  This can work if the total | 
 |      memory requirement for the next TLS block is smaller than the | 
 |      gap.  */ | 
 |  | 
 | #if TLS_TCB_AT_TP | 
 |   /* We simply start with zero.  */ | 
 |   size_t offset = 0; | 
 |  | 
 |   for (size_t cnt = 0; slotinfo[cnt].map != NULL; ++cnt) | 
 |     { | 
 |       assert (cnt < GL(dl_tls_dtv_slotinfo_list)->len); | 
 |  | 
 |       size_t firstbyte = (-slotinfo[cnt].map->l_tls_firstbyte_offset | 
 | 			  & (slotinfo[cnt].map->l_tls_align - 1)); | 
 |       size_t off; | 
 |       max_align = MAX (max_align, slotinfo[cnt].map->l_tls_align); | 
 |  | 
 |       if (freebottom - freetop >= slotinfo[cnt].map->l_tls_blocksize) | 
 | 	{ | 
 | 	  off = roundup (freetop + slotinfo[cnt].map->l_tls_blocksize | 
 | 			 - firstbyte, slotinfo[cnt].map->l_tls_align) | 
 | 		+ firstbyte; | 
 | 	  if (off <= freebottom) | 
 | 	    { | 
 | 	      freetop = off; | 
 |  | 
 | 	      /* XXX For some architectures we perhaps should store the | 
 | 		 negative offset.  */ | 
 | 	      slotinfo[cnt].map->l_tls_offset = off; | 
 | 	      continue; | 
 | 	    } | 
 | 	} | 
 |  | 
 |       off = roundup (offset + slotinfo[cnt].map->l_tls_blocksize - firstbyte, | 
 | 		     slotinfo[cnt].map->l_tls_align) + firstbyte; | 
 |       if (off > offset + slotinfo[cnt].map->l_tls_blocksize | 
 | 		+ (freebottom - freetop)) | 
 | 	{ | 
 | 	  freetop = offset; | 
 | 	  freebottom = off - slotinfo[cnt].map->l_tls_blocksize; | 
 | 	} | 
 |       offset = off; | 
 |  | 
 |       /* XXX For some architectures we perhaps should store the | 
 | 	 negative offset.  */ | 
 |       slotinfo[cnt].map->l_tls_offset = off; | 
 |     } | 
 |  | 
 |   GL(dl_tls_static_used) = offset; | 
 |   GL(dl_tls_static_size) = (roundup (offset + TLS_STATIC_SURPLUS, max_align) | 
 | 			    + TLS_TCB_SIZE); | 
 | #elif TLS_DTV_AT_TP | 
 |   /* The TLS blocks start right after the TCB.  */ | 
 |   size_t offset = TLS_TCB_SIZE; | 
 |  | 
 |   for (size_t cnt = 0; slotinfo[cnt].map != NULL; ++cnt) | 
 |     { | 
 |       assert (cnt < GL(dl_tls_dtv_slotinfo_list)->len); | 
 |  | 
 |       size_t firstbyte = (-slotinfo[cnt].map->l_tls_firstbyte_offset | 
 | 			  & (slotinfo[cnt].map->l_tls_align - 1)); | 
 |       size_t off; | 
 |       max_align = MAX (max_align, slotinfo[cnt].map->l_tls_align); | 
 |  | 
 |       if (slotinfo[cnt].map->l_tls_blocksize <= freetop - freebottom) | 
 | 	{ | 
 | 	  off = roundup (freebottom, slotinfo[cnt].map->l_tls_align); | 
 | 	  if (off - freebottom < firstbyte) | 
 | 	    off += slotinfo[cnt].map->l_tls_align; | 
 | 	  if (off + slotinfo[cnt].map->l_tls_blocksize - firstbyte <= freetop) | 
 | 	    { | 
 | 	      slotinfo[cnt].map->l_tls_offset = off - firstbyte; | 
 | 	      freebottom = (off + slotinfo[cnt].map->l_tls_blocksize | 
 | 			    - firstbyte); | 
 | 	      continue; | 
 | 	    } | 
 | 	} | 
 |  | 
 |       off = roundup (offset, slotinfo[cnt].map->l_tls_align); | 
 |       if (off - offset < firstbyte) | 
 | 	off += slotinfo[cnt].map->l_tls_align; | 
 |  | 
 |       slotinfo[cnt].map->l_tls_offset = off - firstbyte; | 
 |       if (off - firstbyte - offset > freetop - freebottom) | 
 | 	{ | 
 | 	  freebottom = offset; | 
 | 	  freetop = off - firstbyte; | 
 | 	} | 
 |  | 
 |       offset = off + slotinfo[cnt].map->l_tls_blocksize - firstbyte; | 
 |     } | 
 |  | 
 |   GL(dl_tls_static_used) = offset; | 
 |   GL(dl_tls_static_size) = roundup (offset + TLS_STATIC_SURPLUS, | 
 | 				    TLS_TCB_ALIGN); | 
 | #else | 
 | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
 | #endif | 
 |  | 
 |   /* The alignment requirement for the static TLS block.  */ | 
 |   GL(dl_tls_static_align) = max_align; | 
 | } | 
 |  | 
 |  | 
 | /* This is called only when the data structure setup was skipped at startup, | 
 |    when there was no need for it then.  Now we have dynamically loaded | 
 |    something needing TLS, or libpthread needs it.  */ | 
 | int | 
 | internal_function | 
 | _dl_tls_setup (void) | 
 | { | 
 |   assert (GL(dl_tls_dtv_slotinfo_list) == NULL); | 
 |   assert (GL(dl_tls_max_dtv_idx) == 0); | 
 |  | 
 |   const size_t nelem = 2 + TLS_SLOTINFO_SURPLUS; | 
 |  | 
 |   GL(dl_tls_dtv_slotinfo_list) | 
 |     = calloc (1, (sizeof (struct dtv_slotinfo_list) | 
 | 		  + nelem * sizeof (struct dtv_slotinfo))); | 
 |   if (GL(dl_tls_dtv_slotinfo_list) == NULL) | 
 |     return -1; | 
 |  | 
 |   GL(dl_tls_dtv_slotinfo_list)->len = nelem; | 
 |  | 
 |   /* Number of elements in the static TLS block.  It can't be zero | 
 |      because of various assumptions.  The one element is null.  */ | 
 |   GL(dl_tls_static_nelem) = GL(dl_tls_max_dtv_idx) = 1; | 
 |  | 
 |   /* This initializes more variables for us.  */ | 
 |   _dl_determine_tlsoffset (); | 
 |  | 
 |   return 0; | 
 | } | 
 | rtld_hidden_def (_dl_tls_setup) | 
 | #endif | 
 |  | 
 | static void * | 
 | internal_function | 
 | allocate_dtv (void *result) | 
 | { | 
 |   dtv_t *dtv; | 
 |   size_t dtv_length; | 
 |  | 
 |   /* We allocate a few more elements in the dtv than are needed for the | 
 |      initial set of modules.  This should avoid in most cases expansions | 
 |      of the dtv.  */ | 
 |   dtv_length = GL(dl_tls_max_dtv_idx) + DTV_SURPLUS; | 
 |   dtv = calloc (dtv_length + 2, sizeof (dtv_t)); | 
 |   if (dtv != NULL) | 
 |     { | 
 |       /* This is the initial length of the dtv.  */ | 
 |       dtv[0].counter = dtv_length; | 
 |  | 
 |       /* The rest of the dtv (including the generation counter) is | 
 | 	 Initialize with zero to indicate nothing there.  */ | 
 |  | 
 |       /* Add the dtv to the thread data structures.  */ | 
 |       INSTALL_DTV (result, dtv); | 
 |     } | 
 |   else | 
 |     result = NULL; | 
 |  | 
 |   return result; | 
 | } | 
 |  | 
 |  | 
 | /* Get size and alignment requirements of the static TLS block.  */ | 
 | void | 
 | internal_function | 
 | _dl_get_tls_static_info (size_t *sizep, size_t *alignp) | 
 | { | 
 |   *sizep = GL(dl_tls_static_size); | 
 |   *alignp = GL(dl_tls_static_align); | 
 | } | 
 |  | 
 |  | 
 | void * | 
 | internal_function | 
 | _dl_allocate_tls_storage (void) | 
 | { | 
 |   void *result; | 
 |   size_t size = GL(dl_tls_static_size); | 
 |  | 
 | #if TLS_DTV_AT_TP | 
 |   /* Memory layout is: | 
 |      [ TLS_PRE_TCB_SIZE ] [ TLS_TCB_SIZE ] [ TLS blocks ] | 
 | 			  ^ This should be returned.  */ | 
 |   size += (TLS_PRE_TCB_SIZE + GL(dl_tls_static_align) - 1) | 
 | 	  & ~(GL(dl_tls_static_align) - 1); | 
 | #endif | 
 |  | 
 |   /* Allocate a correctly aligned chunk of memory.  */ | 
 |   result = __libc_memalign (GL(dl_tls_static_align), size); | 
 |   if (__builtin_expect (result != NULL, 1)) | 
 |     { | 
 |       /* Allocate the DTV.  */ | 
 |       void *allocated = result; | 
 |  | 
 | #if TLS_TCB_AT_TP | 
 |       /* The TCB follows the TLS blocks.  */ | 
 |       result = (char *) result + size - TLS_TCB_SIZE; | 
 |  | 
 |       /* Clear the TCB data structure.  We can't ask the caller (i.e. | 
 | 	 libpthread) to do it, because we will initialize the DTV et al.  */ | 
 |       memset (result, '\0', TLS_TCB_SIZE); | 
 | #elif TLS_DTV_AT_TP | 
 |       result = (char *) result + size - GL(dl_tls_static_size); | 
 |  | 
 |       /* Clear the TCB data structure and TLS_PRE_TCB_SIZE bytes before it. | 
 | 	 We can't ask the caller (i.e. libpthread) to do it, because we will | 
 | 	 initialize the DTV et al.  */ | 
 |       memset ((char *) result - TLS_PRE_TCB_SIZE, '\0', | 
 | 	      TLS_PRE_TCB_SIZE + TLS_TCB_SIZE); | 
 | #endif | 
 |  | 
 |       result = allocate_dtv (result); | 
 |       if (result == NULL) | 
 | 	free (allocated); | 
 |     } | 
 |  | 
 |   return result; | 
 | } | 
 |  | 
 |  | 
 | #ifndef SHARED | 
 | extern dtv_t _dl_static_dtv[]; | 
 | # define _dl_initial_dtv (&_dl_static_dtv[1]) | 
 | #endif | 
 |  | 
 | static dtv_t * | 
 | _dl_resize_dtv (dtv_t *dtv) | 
 | { | 
 |   /* Resize the dtv.  */ | 
 |   dtv_t *newp; | 
 |   /* Load GL(dl_tls_max_dtv_idx) atomically since it may be written to by | 
 |      other threads concurrently.  */ | 
 |   size_t newsize | 
 |     = atomic_load_acquire (&GL(dl_tls_max_dtv_idx)) + DTV_SURPLUS; | 
 |   size_t oldsize = dtv[-1].counter; | 
 |  | 
 |   if (dtv == GL(dl_initial_dtv)) | 
 |     { | 
 |       /* This is the initial dtv that was either statically allocated in | 
 | 	 __libc_setup_tls or allocated during rtld startup using the | 
 | 	 dl-minimal.c malloc instead of the real malloc.  We can't free | 
 | 	 it, we have to abandon the old storage.  */ | 
 |  | 
 |       newp = malloc ((2 + newsize) * sizeof (dtv_t)); | 
 |       if (newp == NULL) | 
 | 	oom (); | 
 |       memcpy (newp, &dtv[-1], (2 + oldsize) * sizeof (dtv_t)); | 
 |     } | 
 |   else | 
 |     { | 
 |       newp = realloc (&dtv[-1], | 
 | 		      (2 + newsize) * sizeof (dtv_t)); | 
 |       if (newp == NULL) | 
 | 	oom (); | 
 |     } | 
 |  | 
 |   newp[0].counter = newsize; | 
 |  | 
 |   /* Clear the newly allocated part.  */ | 
 |   memset (newp + 2 + oldsize, '\0', | 
 | 	  (newsize - oldsize) * sizeof (dtv_t)); | 
 |  | 
 |   /* Return the generation counter.  */ | 
 |   return &newp[1]; | 
 | } | 
 |  | 
 |  | 
 | void * | 
 | internal_function | 
 | _dl_allocate_tls_init (void *result) | 
 | { | 
 |   if (result == NULL) | 
 |     /* The memory allocation failed.  */ | 
 |     return NULL; | 
 |  | 
 |   dtv_t *dtv = GET_DTV (result); | 
 |   struct dtv_slotinfo_list *listp; | 
 |   size_t total = 0; | 
 |   size_t maxgen = 0; | 
 |  | 
 |   /* Check if the current dtv is big enough.   */ | 
 |   if (dtv[-1].counter < GL(dl_tls_max_dtv_idx)) | 
 |     { | 
 |       /* Resize the dtv.  */ | 
 |       dtv = _dl_resize_dtv (dtv); | 
 |  | 
 |       /* Install this new dtv in the thread data structures.  */ | 
 |       INSTALL_DTV (result, &dtv[-1]); | 
 |     } | 
 |  | 
 |   /* We have to prepare the dtv for all currently loaded modules using | 
 |      TLS.  For those which are dynamically loaded we add the values | 
 |      indicating deferred allocation.  */ | 
 |   listp = GL(dl_tls_dtv_slotinfo_list); | 
 |   while (1) | 
 |     { | 
 |       size_t cnt; | 
 |  | 
 |       for (cnt = total == 0 ? 1 : 0; cnt < listp->len; ++cnt) | 
 | 	{ | 
 | 	  struct link_map *map; | 
 | 	  void *dest; | 
 |  | 
 | 	  /* Check for the total number of used slots.  */ | 
 | 	  if (total + cnt > GL(dl_tls_max_dtv_idx)) | 
 | 	    break; | 
 |  | 
 | 	  map = listp->slotinfo[cnt].map; | 
 | 	  if (map == NULL) | 
 | 	    /* Unused entry.  */ | 
 | 	    continue; | 
 |  | 
 | 	  /* Keep track of the maximum generation number.  This might | 
 | 	     not be the generation counter.  */ | 
 | 	  assert (listp->slotinfo[cnt].gen <= GL(dl_tls_generation)); | 
 | 	  maxgen = MAX (maxgen, listp->slotinfo[cnt].gen); | 
 |  | 
 | 	  dtv[map->l_tls_modid].pointer.val = TLS_DTV_UNALLOCATED; | 
 | 	  dtv[map->l_tls_modid].pointer.is_static = false; | 
 |  | 
 | 	  if (map->l_tls_offset == NO_TLS_OFFSET | 
 | 	      || map->l_tls_offset == FORCED_DYNAMIC_TLS_OFFSET) | 
 | 	    continue; | 
 |  | 
 | 	  assert (map->l_tls_modid == total + cnt); | 
 | 	  assert (map->l_tls_blocksize >= map->l_tls_initimage_size); | 
 | #if TLS_TCB_AT_TP | 
 | 	  assert ((size_t) map->l_tls_offset >= map->l_tls_blocksize); | 
 | 	  dest = (char *) result - map->l_tls_offset; | 
 | #elif TLS_DTV_AT_TP | 
 | 	  dest = (char *) result + map->l_tls_offset; | 
 | #else | 
 | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
 | #endif | 
 |  | 
 | 	  /* Copy the initialization image and clear the BSS part.  */ | 
 | 	  memset (__mempcpy (dest, map->l_tls_initimage, | 
 | 			     map->l_tls_initimage_size), '\0', | 
 | 		  map->l_tls_blocksize - map->l_tls_initimage_size); | 
 | 	} | 
 |  | 
 |       total += cnt; | 
 |       if (total >= GL(dl_tls_max_dtv_idx)) | 
 | 	break; | 
 |  | 
 |       listp = listp->next; | 
 |       assert (listp != NULL); | 
 |     } | 
 |  | 
 |   /* The DTV version is up-to-date now.  */ | 
 |   dtv[0].counter = maxgen; | 
 |  | 
 |   return result; | 
 | } | 
 | rtld_hidden_def (_dl_allocate_tls_init) | 
 |  | 
 | void * | 
 | internal_function | 
 | _dl_allocate_tls (void *mem) | 
 | { | 
 |   return _dl_allocate_tls_init (mem == NULL | 
 | 				? _dl_allocate_tls_storage () | 
 | 				: allocate_dtv (mem)); | 
 | } | 
 | rtld_hidden_def (_dl_allocate_tls) | 
 |  | 
 |  | 
 | void | 
 | internal_function | 
 | _dl_deallocate_tls (void *tcb, bool dealloc_tcb) | 
 | { | 
 |   dtv_t *dtv = GET_DTV (tcb); | 
 |  | 
 |   /* We need to free the memory allocated for non-static TLS.  */ | 
 |   for (size_t cnt = 0; cnt < dtv[-1].counter; ++cnt) | 
 |     if (! dtv[1 + cnt].pointer.is_static | 
 | 	&& dtv[1 + cnt].pointer.val != TLS_DTV_UNALLOCATED) | 
 |       free (dtv[1 + cnt].pointer.val); | 
 |  | 
 |   /* The array starts with dtv[-1].  */ | 
 |   if (dtv != GL(dl_initial_dtv)) | 
 |     free (dtv - 1); | 
 |  | 
 |   if (dealloc_tcb) | 
 |     { | 
 | #if TLS_TCB_AT_TP | 
 |       /* The TCB follows the TLS blocks.  Back up to free the whole block.  */ | 
 |       tcb -= GL(dl_tls_static_size) - TLS_TCB_SIZE; | 
 | #elif TLS_DTV_AT_TP | 
 |       /* Back up the TLS_PRE_TCB_SIZE bytes.  */ | 
 |       tcb -= (TLS_PRE_TCB_SIZE + GL(dl_tls_static_align) - 1) | 
 | 	     & ~(GL(dl_tls_static_align) - 1); | 
 | #endif | 
 |       free (tcb); | 
 |     } | 
 | } | 
 | rtld_hidden_def (_dl_deallocate_tls) | 
 |  | 
 |  | 
 | #ifdef SHARED | 
 | /* The __tls_get_addr function has two basic forms which differ in the | 
 |    arguments.  The IA-64 form takes two parameters, the module ID and | 
 |    offset.  The form used, among others, on IA-32 takes a reference to | 
 |    a special structure which contain the same information.  The second | 
 |    form seems to be more often used (in the moment) so we default to | 
 |    it.  Users of the IA-64 form have to provide adequate definitions | 
 |    of the following macros.  */ | 
 | # ifndef GET_ADDR_ARGS | 
 | #  define GET_ADDR_ARGS tls_index *ti | 
 | #  define GET_ADDR_PARAM ti | 
 | # endif | 
 | # ifndef GET_ADDR_MODULE | 
 | #  define GET_ADDR_MODULE ti->ti_module | 
 | # endif | 
 | # ifndef GET_ADDR_OFFSET | 
 | #  define GET_ADDR_OFFSET ti->ti_offset | 
 | # endif | 
 |  | 
 |  | 
 | static void * | 
 | allocate_and_init (struct link_map *map) | 
 | { | 
 |   void *newp; | 
 |  | 
 |   newp = __libc_memalign (map->l_tls_align, map->l_tls_blocksize); | 
 |   if (newp == NULL) | 
 |     oom (); | 
 |  | 
 |   /* Initialize the memory.  */ | 
 |   memset (__mempcpy (newp, map->l_tls_initimage, map->l_tls_initimage_size), | 
 | 	  '\0', map->l_tls_blocksize - map->l_tls_initimage_size); | 
 |  | 
 |   return newp; | 
 | } | 
 |  | 
 |  | 
 | struct link_map * | 
 | _dl_update_slotinfo (unsigned long int req_modid) | 
 | { | 
 |   struct link_map *the_map = NULL; | 
 |   dtv_t *dtv = THREAD_DTV (); | 
 |  | 
 |   /* The global dl_tls_dtv_slotinfo array contains for each module | 
 |      index the generation counter current when the entry was created. | 
 |      This array never shrinks so that all module indices which were | 
 |      valid at some time can be used to access it.  Before the first | 
 |      use of a new module index in this function the array was extended | 
 |      appropriately.  Access also does not have to be guarded against | 
 |      modifications of the array.  It is assumed that pointer-size | 
 |      values can be read atomically even in SMP environments.  It is | 
 |      possible that other threads at the same time dynamically load | 
 |      code and therefore add to the slotinfo list.  This is a problem | 
 |      since we must not pick up any information about incomplete work. | 
 |      The solution to this is to ignore all dtv slots which were | 
 |      created after the one we are currently interested.  We know that | 
 |      dynamic loading for this module is completed and this is the last | 
 |      load operation we know finished.  */ | 
 |   unsigned long int idx = req_modid; | 
 |   struct dtv_slotinfo_list *listp = GL(dl_tls_dtv_slotinfo_list); | 
 |  | 
 |   while (idx >= listp->len) | 
 |     { | 
 |       idx -= listp->len; | 
 |       listp = listp->next; | 
 |     } | 
 |  | 
 |   if (dtv[0].counter < listp->slotinfo[idx].gen) | 
 |     { | 
 |       /* The generation counter for the slot is higher than what the | 
 | 	 current dtv implements.  We have to update the whole dtv but | 
 | 	 only those entries with a generation counter <= the one for | 
 | 	 the entry we need.  */ | 
 |       size_t new_gen = listp->slotinfo[idx].gen; | 
 |       size_t total = 0; | 
 |  | 
 |       /* We have to look through the entire dtv slotinfo list.  */ | 
 |       listp =  GL(dl_tls_dtv_slotinfo_list); | 
 |       do | 
 | 	{ | 
 | 	  for (size_t cnt = total == 0 ? 1 : 0; cnt < listp->len; ++cnt) | 
 | 	    { | 
 | 	      size_t gen = listp->slotinfo[cnt].gen; | 
 |  | 
 | 	      if (gen > new_gen) | 
 | 		/* This is a slot for a generation younger than the | 
 | 		   one we are handling now.  It might be incompletely | 
 | 		   set up so ignore it.  */ | 
 | 		continue; | 
 |  | 
 | 	      /* If the entry is older than the current dtv layout we | 
 | 		 know we don't have to handle it.  */ | 
 | 	      if (gen <= dtv[0].counter) | 
 | 		continue; | 
 |  | 
 | 	      /* If there is no map this means the entry is empty.  */ | 
 | 	      struct link_map *map = listp->slotinfo[cnt].map; | 
 | 	      if (map == NULL) | 
 | 		{ | 
 | 		  if (dtv[-1].counter >= total + cnt) | 
 | 		    { | 
 | 		      /* If this modid was used at some point the memory | 
 | 			 might still be allocated.  */ | 
 | 		      if (! dtv[total + cnt].pointer.is_static | 
 | 			  && (dtv[total + cnt].pointer.val | 
 | 			      != TLS_DTV_UNALLOCATED)) | 
 | 			free (dtv[total + cnt].pointer.val); | 
 | 		      dtv[total + cnt].pointer.val = TLS_DTV_UNALLOCATED; | 
 | 		      dtv[total + cnt].pointer.is_static = false; | 
 | 		    } | 
 |  | 
 | 		  continue; | 
 | 		} | 
 |  | 
 | 	      /* Check whether the current dtv array is large enough.  */ | 
 | 	      size_t modid = map->l_tls_modid; | 
 | 	      assert (total + cnt == modid); | 
 | 	      if (dtv[-1].counter < modid) | 
 | 		{ | 
 | 		  /* Resize the dtv.  */ | 
 | 		  dtv = _dl_resize_dtv (dtv); | 
 |  | 
 | 		  assert (modid <= dtv[-1].counter); | 
 |  | 
 | 		  /* Install this new dtv in the thread data | 
 | 		     structures.  */ | 
 | 		  INSTALL_NEW_DTV (dtv); | 
 | 		} | 
 |  | 
 | 	      /* If there is currently memory allocate for this | 
 | 		 dtv entry free it.  */ | 
 | 	      /* XXX Ideally we will at some point create a memory | 
 | 		 pool.  */ | 
 | 	      if (! dtv[modid].pointer.is_static | 
 | 		  && dtv[modid].pointer.val != TLS_DTV_UNALLOCATED) | 
 | 		/* Note that free is called for NULL is well.  We | 
 | 		   deallocate even if it is this dtv entry we are | 
 | 		   supposed to load.  The reason is that we call | 
 | 		   memalign and not malloc.  */ | 
 | 		free (dtv[modid].pointer.val); | 
 |  | 
 | 	      dtv[modid].pointer.val = TLS_DTV_UNALLOCATED; | 
 | 	      dtv[modid].pointer.is_static = false; | 
 |  | 
 | 	      if (modid == req_modid) | 
 | 		the_map = map; | 
 | 	    } | 
 |  | 
 | 	  total += listp->len; | 
 | 	} | 
 |       while ((listp = listp->next) != NULL); | 
 |  | 
 |       /* This will be the new maximum generation counter.  */ | 
 |       dtv[0].counter = new_gen; | 
 |     } | 
 |  | 
 |   return the_map; | 
 | } | 
 |  | 
 |  | 
 | static void * | 
 | __attribute_noinline__ | 
 | tls_get_addr_tail (GET_ADDR_ARGS, dtv_t *dtv, struct link_map *the_map) | 
 | { | 
 |   /* The allocation was deferred.  Do it now.  */ | 
 |   if (the_map == NULL) | 
 |     { | 
 |       /* Find the link map for this module.  */ | 
 |       size_t idx = GET_ADDR_MODULE; | 
 |       struct dtv_slotinfo_list *listp = GL(dl_tls_dtv_slotinfo_list); | 
 |  | 
 |       while (idx >= listp->len) | 
 | 	{ | 
 | 	  idx -= listp->len; | 
 | 	  listp = listp->next; | 
 | 	} | 
 |  | 
 |       the_map = listp->slotinfo[idx].map; | 
 |     } | 
 |  | 
 |   /* Make sure that, if a dlopen running in parallel forces the | 
 |      variable into static storage, we'll wait until the address in the | 
 |      static TLS block is set up, and use that.  If we're undecided | 
 |      yet, make sure we make the decision holding the lock as well.  */ | 
 |   if (__glibc_unlikely (the_map->l_tls_offset | 
 | 			!= FORCED_DYNAMIC_TLS_OFFSET)) | 
 |     { | 
 |       __rtld_lock_lock_recursive (GL(dl_load_lock)); | 
 |       if (__glibc_likely (the_map->l_tls_offset == NO_TLS_OFFSET)) | 
 | 	{ | 
 | 	  the_map->l_tls_offset = FORCED_DYNAMIC_TLS_OFFSET; | 
 | 	  __rtld_lock_unlock_recursive (GL(dl_load_lock)); | 
 | 	} | 
 |       else if (__glibc_likely (the_map->l_tls_offset | 
 | 			       != FORCED_DYNAMIC_TLS_OFFSET)) | 
 | 	{ | 
 | #if TLS_TCB_AT_TP | 
 | 	  void *p = (char *) THREAD_SELF - the_map->l_tls_offset; | 
 | #elif TLS_DTV_AT_TP | 
 | 	  void *p = (char *) THREAD_SELF + the_map->l_tls_offset + TLS_PRE_TCB_SIZE; | 
 | #else | 
 | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
 | #endif | 
 | 	  __rtld_lock_unlock_recursive (GL(dl_load_lock)); | 
 |  | 
 | 	  dtv[GET_ADDR_MODULE].pointer.is_static = true; | 
 | 	  dtv[GET_ADDR_MODULE].pointer.val = p; | 
 |  | 
 | 	  return (char *) p + GET_ADDR_OFFSET; | 
 | 	} | 
 |       else | 
 | 	__rtld_lock_unlock_recursive (GL(dl_load_lock)); | 
 |     } | 
 |   void *p = dtv[GET_ADDR_MODULE].pointer.val = allocate_and_init (the_map); | 
 |   assert (!dtv[GET_ADDR_MODULE].pointer.is_static); | 
 |  | 
 |   return (char *) p + GET_ADDR_OFFSET; | 
 | } | 
 |  | 
 |  | 
 | static struct link_map * | 
 | __attribute_noinline__ | 
 | update_get_addr (GET_ADDR_ARGS) | 
 | { | 
 |   struct link_map *the_map = _dl_update_slotinfo (GET_ADDR_MODULE); | 
 |   dtv_t *dtv = THREAD_DTV (); | 
 |  | 
 |   void *p = dtv[GET_ADDR_MODULE].pointer.val; | 
 |  | 
 |   if (__glibc_unlikely (p == TLS_DTV_UNALLOCATED)) | 
 |     return tls_get_addr_tail (GET_ADDR_PARAM, dtv, the_map); | 
 |  | 
 |   return (void *) p + GET_ADDR_OFFSET; | 
 | } | 
 |  | 
 | /* For all machines that have a non-macro version of __tls_get_addr, we | 
 |    want to use rtld_hidden_proto/rtld_hidden_def in order to call the | 
 |    internal alias for __tls_get_addr from ld.so. This avoids a PLT entry | 
 |    in ld.so for __tls_get_addr.  */ | 
 |  | 
 | #ifndef __tls_get_addr | 
 | extern void * __tls_get_addr (GET_ADDR_ARGS); | 
 | rtld_hidden_proto (__tls_get_addr) | 
 | rtld_hidden_def (__tls_get_addr) | 
 | #endif | 
 |  | 
 | /* The generic dynamic and local dynamic model cannot be used in | 
 |    statically linked applications.  */ | 
 | void * | 
 | __tls_get_addr (GET_ADDR_ARGS) | 
 | { | 
 |   dtv_t *dtv = THREAD_DTV (); | 
 |  | 
 |   if (__glibc_unlikely (dtv[0].counter != GL(dl_tls_generation))) | 
 |     return update_get_addr (GET_ADDR_PARAM); | 
 |  | 
 |   void *p = dtv[GET_ADDR_MODULE].pointer.val; | 
 |  | 
 |   if (__glibc_unlikely (p == TLS_DTV_UNALLOCATED)) | 
 |     return tls_get_addr_tail (GET_ADDR_PARAM, dtv, NULL); | 
 |  | 
 |   return (char *) p + GET_ADDR_OFFSET; | 
 | } | 
 | #endif | 
 |  | 
 |  | 
 | /* Look up the module's TLS block as for __tls_get_addr, | 
 |    but never touch anything.  Return null if it's not allocated yet.  */ | 
 | void * | 
 | _dl_tls_get_addr_soft (struct link_map *l) | 
 | { | 
 |   if (__glibc_unlikely (l->l_tls_modid == 0)) | 
 |     /* This module has no TLS segment.  */ | 
 |     return NULL; | 
 |  | 
 |   dtv_t *dtv = THREAD_DTV (); | 
 |   if (__glibc_unlikely (dtv[0].counter != GL(dl_tls_generation))) | 
 |     { | 
 |       /* This thread's DTV is not completely current, | 
 | 	 but it might already cover this module.  */ | 
 |  | 
 |       if (l->l_tls_modid >= dtv[-1].counter) | 
 | 	/* Nope.  */ | 
 | 	return NULL; | 
 |  | 
 |       size_t idx = l->l_tls_modid; | 
 |       struct dtv_slotinfo_list *listp = GL(dl_tls_dtv_slotinfo_list); | 
 |       while (idx >= listp->len) | 
 | 	{ | 
 | 	  idx -= listp->len; | 
 | 	  listp = listp->next; | 
 | 	} | 
 |  | 
 |       /* We've reached the slot for this module. | 
 | 	 If its generation counter is higher than the DTV's, | 
 | 	 this thread does not know about this module yet.  */ | 
 |       if (dtv[0].counter < listp->slotinfo[idx].gen) | 
 | 	return NULL; | 
 |     } | 
 |  | 
 |   void *data = dtv[l->l_tls_modid].pointer.val; | 
 |   if (__glibc_unlikely (data == TLS_DTV_UNALLOCATED)) | 
 |     /* The DTV is current, but this thread has not yet needed | 
 |        to allocate this module's segment.  */ | 
 |     data = NULL; | 
 |  | 
 |   return data; | 
 | } | 
 |  | 
 |  | 
 | void | 
 | _dl_add_to_slotinfo (struct link_map *l) | 
 | { | 
 |   /* Now that we know the object is loaded successfully add | 
 |      modules containing TLS data to the dtv info table.  We | 
 |      might have to increase its size.  */ | 
 |   struct dtv_slotinfo_list *listp; | 
 |   struct dtv_slotinfo_list *prevp; | 
 |   size_t idx = l->l_tls_modid; | 
 |  | 
 |   /* Find the place in the dtv slotinfo list.  */ | 
 |   listp = GL(dl_tls_dtv_slotinfo_list); | 
 |   prevp = NULL;		/* Needed to shut up gcc.  */ | 
 |   do | 
 |     { | 
 |       /* Does it fit in the array of this list element?  */ | 
 |       if (idx < listp->len) | 
 | 	break; | 
 |       idx -= listp->len; | 
 |       prevp = listp; | 
 |       listp = listp->next; | 
 |     } | 
 |   while (listp != NULL); | 
 |  | 
 |   if (listp == NULL) | 
 |     { | 
 |       /* When we come here it means we have to add a new element | 
 | 	 to the slotinfo list.  And the new module must be in | 
 | 	 the first slot.  */ | 
 |       assert (idx == 0); | 
 |  | 
 |       listp = prevp->next = (struct dtv_slotinfo_list *) | 
 | 	malloc (sizeof (struct dtv_slotinfo_list) | 
 | 		+ TLS_SLOTINFO_SURPLUS * sizeof (struct dtv_slotinfo)); | 
 |       if (listp == NULL) | 
 | 	{ | 
 | 	  /* We ran out of memory.  We will simply fail this | 
 | 	     call but don't undo anything we did so far.  The | 
 | 	     application will crash or be terminated anyway very | 
 | 	     soon.  */ | 
 |  | 
 | 	  /* We have to do this since some entries in the dtv | 
 | 	     slotinfo array might already point to this | 
 | 	     generation.  */ | 
 | 	  ++GL(dl_tls_generation); | 
 |  | 
 | 	  _dl_signal_error (ENOMEM, "dlopen", NULL, N_("\ | 
 | cannot create TLS data structures")); | 
 | 	} | 
 |  | 
 |       listp->len = TLS_SLOTINFO_SURPLUS; | 
 |       listp->next = NULL; | 
 |       memset (listp->slotinfo, '\0', | 
 | 	      TLS_SLOTINFO_SURPLUS * sizeof (struct dtv_slotinfo)); | 
 |     } | 
 |  | 
 |   /* Add the information into the slotinfo data structure.  */ | 
 |   listp->slotinfo[idx].map = l; | 
 |   listp->slotinfo[idx].gen = GL(dl_tls_generation) + 1; | 
 | } |