|  | /* Thread-local storage handling in the ELF dynamic linker.  Generic version. | 
|  | Copyright (C) 2002-2016 Free Software Foundation, Inc. | 
|  | This file is part of the GNU C Library. | 
|  |  | 
|  | The GNU C Library is free software; you can redistribute it and/or | 
|  | modify it under the terms of the GNU Lesser General Public | 
|  | License as published by the Free Software Foundation; either | 
|  | version 2.1 of the License, or (at your option) any later version. | 
|  |  | 
|  | The GNU C Library is distributed in the hope that it will be useful, | 
|  | but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|  | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU | 
|  | Lesser General Public License for more details. | 
|  |  | 
|  | You should have received a copy of the GNU Lesser General Public | 
|  | License along with the GNU C Library; if not, see | 
|  | <http://www.gnu.org/licenses/>.  */ | 
|  |  | 
|  | #include <assert.h> | 
|  | #include <errno.h> | 
|  | #include <libintl.h> | 
|  | #include <signal.h> | 
|  | #include <stdlib.h> | 
|  | #include <unistd.h> | 
|  | #include <sys/param.h> | 
|  | #include <atomic.h> | 
|  |  | 
|  | #include <tls.h> | 
|  | #include <dl-tls.h> | 
|  | #include <ldsodefs.h> | 
|  |  | 
|  | /* Amount of excess space to allocate in the static TLS area | 
|  | to allow dynamic loading of modules defining IE-model TLS data.  */ | 
|  | #define TLS_STATIC_SURPLUS	64 + DL_NNS * 100 | 
|  |  | 
|  |  | 
|  | /* Out-of-memory handler.  */ | 
|  | static void | 
|  | __attribute__ ((__noreturn__)) | 
|  | oom (void) | 
|  | { | 
|  | _dl_fatal_printf ("cannot allocate memory for thread-local data: ABORT\n"); | 
|  | } | 
|  |  | 
|  |  | 
|  | size_t | 
|  | internal_function | 
|  | _dl_next_tls_modid (void) | 
|  | { | 
|  | size_t result; | 
|  |  | 
|  | if (__builtin_expect (GL(dl_tls_dtv_gaps), false)) | 
|  | { | 
|  | size_t disp = 0; | 
|  | struct dtv_slotinfo_list *runp = GL(dl_tls_dtv_slotinfo_list); | 
|  |  | 
|  | /* Note that this branch will never be executed during program | 
|  | start since there are no gaps at that time.  Therefore it | 
|  | does not matter that the dl_tls_dtv_slotinfo is not allocated | 
|  | yet when the function is called for the first times. | 
|  |  | 
|  | NB: the offset +1 is due to the fact that DTV[0] is used | 
|  | for something else.  */ | 
|  | result = GL(dl_tls_static_nelem) + 1; | 
|  | if (result <= GL(dl_tls_max_dtv_idx)) | 
|  | do | 
|  | { | 
|  | while (result - disp < runp->len) | 
|  | { | 
|  | if (runp->slotinfo[result - disp].map == NULL) | 
|  | break; | 
|  |  | 
|  | ++result; | 
|  | assert (result <= GL(dl_tls_max_dtv_idx) + 1); | 
|  | } | 
|  |  | 
|  | if (result - disp < runp->len) | 
|  | break; | 
|  |  | 
|  | disp += runp->len; | 
|  | } | 
|  | while ((runp = runp->next) != NULL); | 
|  |  | 
|  | if (result > GL(dl_tls_max_dtv_idx)) | 
|  | { | 
|  | /* The new index must indeed be exactly one higher than the | 
|  | previous high.  */ | 
|  | assert (result == GL(dl_tls_max_dtv_idx) + 1); | 
|  | /* There is no gap anymore.  */ | 
|  | GL(dl_tls_dtv_gaps) = false; | 
|  |  | 
|  | goto nogaps; | 
|  | } | 
|  | } | 
|  | else | 
|  | { | 
|  | /* No gaps, allocate a new entry.  */ | 
|  | nogaps: | 
|  |  | 
|  | result = ++GL(dl_tls_max_dtv_idx); | 
|  | } | 
|  |  | 
|  | return result; | 
|  | } | 
|  |  | 
|  |  | 
|  | size_t | 
|  | internal_function | 
|  | _dl_count_modids (void) | 
|  | { | 
|  | /* It is rare that we have gaps; see elf/dl-open.c (_dl_open) where | 
|  | we fail to load a module and unload it leaving a gap.  If we don't | 
|  | have gaps then the number of modids is the current maximum so | 
|  | return that.  */ | 
|  | if (__glibc_likely (!GL(dl_tls_dtv_gaps))) | 
|  | return GL(dl_tls_max_dtv_idx); | 
|  |  | 
|  | /* We have gaps and are forced to count the non-NULL entries.  */ | 
|  | size_t n = 0; | 
|  | struct dtv_slotinfo_list *runp = GL(dl_tls_dtv_slotinfo_list); | 
|  | while (runp != NULL) | 
|  | { | 
|  | for (size_t i = 0; i < runp->len; ++i) | 
|  | if (runp->slotinfo[i].map != NULL) | 
|  | ++n; | 
|  |  | 
|  | runp = runp->next; | 
|  | } | 
|  |  | 
|  | return n; | 
|  | } | 
|  |  | 
|  |  | 
|  | #ifdef SHARED | 
|  | void | 
|  | internal_function | 
|  | _dl_determine_tlsoffset (void) | 
|  | { | 
|  | size_t max_align = TLS_TCB_ALIGN; | 
|  | size_t freetop = 0; | 
|  | size_t freebottom = 0; | 
|  |  | 
|  | /* The first element of the dtv slot info list is allocated.  */ | 
|  | assert (GL(dl_tls_dtv_slotinfo_list) != NULL); | 
|  | /* There is at this point only one element in the | 
|  | dl_tls_dtv_slotinfo_list list.  */ | 
|  | assert (GL(dl_tls_dtv_slotinfo_list)->next == NULL); | 
|  |  | 
|  | struct dtv_slotinfo *slotinfo = GL(dl_tls_dtv_slotinfo_list)->slotinfo; | 
|  |  | 
|  | /* Determining the offset of the various parts of the static TLS | 
|  | block has several dependencies.  In addition we have to work | 
|  | around bugs in some toolchains. | 
|  |  | 
|  | Each TLS block from the objects available at link time has a size | 
|  | and an alignment requirement.  The GNU ld computes the alignment | 
|  | requirements for the data at the positions *in the file*, though. | 
|  | I.e, it is not simply possible to allocate a block with the size | 
|  | of the TLS program header entry.  The data is layed out assuming | 
|  | that the first byte of the TLS block fulfills | 
|  |  | 
|  | p_vaddr mod p_align == &TLS_BLOCK mod p_align | 
|  |  | 
|  | This means we have to add artificial padding at the beginning of | 
|  | the TLS block.  These bytes are never used for the TLS data in | 
|  | this module but the first byte allocated must be aligned | 
|  | according to mod p_align == 0 so that the first byte of the TLS | 
|  | block is aligned according to p_vaddr mod p_align.  This is ugly | 
|  | and the linker can help by computing the offsets in the TLS block | 
|  | assuming the first byte of the TLS block is aligned according to | 
|  | p_align. | 
|  |  | 
|  | The extra space which might be allocated before the first byte of | 
|  | the TLS block need not go unused.  The code below tries to use | 
|  | that memory for the next TLS block.  This can work if the total | 
|  | memory requirement for the next TLS block is smaller than the | 
|  | gap.  */ | 
|  |  | 
|  | #if TLS_TCB_AT_TP | 
|  | /* We simply start with zero.  */ | 
|  | size_t offset = 0; | 
|  |  | 
|  | for (size_t cnt = 0; slotinfo[cnt].map != NULL; ++cnt) | 
|  | { | 
|  | assert (cnt < GL(dl_tls_dtv_slotinfo_list)->len); | 
|  |  | 
|  | size_t firstbyte = (-slotinfo[cnt].map->l_tls_firstbyte_offset | 
|  | & (slotinfo[cnt].map->l_tls_align - 1)); | 
|  | size_t off; | 
|  | max_align = MAX (max_align, slotinfo[cnt].map->l_tls_align); | 
|  |  | 
|  | if (freebottom - freetop >= slotinfo[cnt].map->l_tls_blocksize) | 
|  | { | 
|  | off = roundup (freetop + slotinfo[cnt].map->l_tls_blocksize | 
|  | - firstbyte, slotinfo[cnt].map->l_tls_align) | 
|  | + firstbyte; | 
|  | if (off <= freebottom) | 
|  | { | 
|  | freetop = off; | 
|  |  | 
|  | /* XXX For some architectures we perhaps should store the | 
|  | negative offset.  */ | 
|  | slotinfo[cnt].map->l_tls_offset = off; | 
|  | continue; | 
|  | } | 
|  | } | 
|  |  | 
|  | off = roundup (offset + slotinfo[cnt].map->l_tls_blocksize - firstbyte, | 
|  | slotinfo[cnt].map->l_tls_align) + firstbyte; | 
|  | if (off > offset + slotinfo[cnt].map->l_tls_blocksize | 
|  | + (freebottom - freetop)) | 
|  | { | 
|  | freetop = offset; | 
|  | freebottom = off - slotinfo[cnt].map->l_tls_blocksize; | 
|  | } | 
|  | offset = off; | 
|  |  | 
|  | /* XXX For some architectures we perhaps should store the | 
|  | negative offset.  */ | 
|  | slotinfo[cnt].map->l_tls_offset = off; | 
|  | } | 
|  |  | 
|  | GL(dl_tls_static_used) = offset; | 
|  | GL(dl_tls_static_size) = (roundup (offset + TLS_STATIC_SURPLUS, max_align) | 
|  | + TLS_TCB_SIZE); | 
|  | #elif TLS_DTV_AT_TP | 
|  | /* The TLS blocks start right after the TCB.  */ | 
|  | size_t offset = TLS_TCB_SIZE; | 
|  |  | 
|  | for (size_t cnt = 0; slotinfo[cnt].map != NULL; ++cnt) | 
|  | { | 
|  | assert (cnt < GL(dl_tls_dtv_slotinfo_list)->len); | 
|  |  | 
|  | size_t firstbyte = (-slotinfo[cnt].map->l_tls_firstbyte_offset | 
|  | & (slotinfo[cnt].map->l_tls_align - 1)); | 
|  | size_t off; | 
|  | max_align = MAX (max_align, slotinfo[cnt].map->l_tls_align); | 
|  |  | 
|  | if (slotinfo[cnt].map->l_tls_blocksize <= freetop - freebottom) | 
|  | { | 
|  | off = roundup (freebottom, slotinfo[cnt].map->l_tls_align); | 
|  | if (off - freebottom < firstbyte) | 
|  | off += slotinfo[cnt].map->l_tls_align; | 
|  | if (off + slotinfo[cnt].map->l_tls_blocksize - firstbyte <= freetop) | 
|  | { | 
|  | slotinfo[cnt].map->l_tls_offset = off - firstbyte; | 
|  | freebottom = (off + slotinfo[cnt].map->l_tls_blocksize | 
|  | - firstbyte); | 
|  | continue; | 
|  | } | 
|  | } | 
|  |  | 
|  | off = roundup (offset, slotinfo[cnt].map->l_tls_align); | 
|  | if (off - offset < firstbyte) | 
|  | off += slotinfo[cnt].map->l_tls_align; | 
|  |  | 
|  | slotinfo[cnt].map->l_tls_offset = off - firstbyte; | 
|  | if (off - firstbyte - offset > freetop - freebottom) | 
|  | { | 
|  | freebottom = offset; | 
|  | freetop = off - firstbyte; | 
|  | } | 
|  |  | 
|  | offset = off + slotinfo[cnt].map->l_tls_blocksize - firstbyte; | 
|  | } | 
|  |  | 
|  | GL(dl_tls_static_used) = offset; | 
|  | GL(dl_tls_static_size) = roundup (offset + TLS_STATIC_SURPLUS, | 
|  | TLS_TCB_ALIGN); | 
|  | #else | 
|  | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
|  | #endif | 
|  |  | 
|  | /* The alignment requirement for the static TLS block.  */ | 
|  | GL(dl_tls_static_align) = max_align; | 
|  | } | 
|  |  | 
|  |  | 
|  | /* This is called only when the data structure setup was skipped at startup, | 
|  | when there was no need for it then.  Now we have dynamically loaded | 
|  | something needing TLS, or libpthread needs it.  */ | 
|  | int | 
|  | internal_function | 
|  | _dl_tls_setup (void) | 
|  | { | 
|  | assert (GL(dl_tls_dtv_slotinfo_list) == NULL); | 
|  | assert (GL(dl_tls_max_dtv_idx) == 0); | 
|  |  | 
|  | const size_t nelem = 2 + TLS_SLOTINFO_SURPLUS; | 
|  |  | 
|  | GL(dl_tls_dtv_slotinfo_list) | 
|  | = calloc (1, (sizeof (struct dtv_slotinfo_list) | 
|  | + nelem * sizeof (struct dtv_slotinfo))); | 
|  | if (GL(dl_tls_dtv_slotinfo_list) == NULL) | 
|  | return -1; | 
|  |  | 
|  | GL(dl_tls_dtv_slotinfo_list)->len = nelem; | 
|  |  | 
|  | /* Number of elements in the static TLS block.  It can't be zero | 
|  | because of various assumptions.  The one element is null.  */ | 
|  | GL(dl_tls_static_nelem) = GL(dl_tls_max_dtv_idx) = 1; | 
|  |  | 
|  | /* This initializes more variables for us.  */ | 
|  | _dl_determine_tlsoffset (); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | rtld_hidden_def (_dl_tls_setup) | 
|  | #endif | 
|  |  | 
|  | static void * | 
|  | internal_function | 
|  | allocate_dtv (void *result) | 
|  | { | 
|  | dtv_t *dtv; | 
|  | size_t dtv_length; | 
|  |  | 
|  | /* We allocate a few more elements in the dtv than are needed for the | 
|  | initial set of modules.  This should avoid in most cases expansions | 
|  | of the dtv.  */ | 
|  | dtv_length = GL(dl_tls_max_dtv_idx) + DTV_SURPLUS; | 
|  | dtv = calloc (dtv_length + 2, sizeof (dtv_t)); | 
|  | if (dtv != NULL) | 
|  | { | 
|  | /* This is the initial length of the dtv.  */ | 
|  | dtv[0].counter = dtv_length; | 
|  |  | 
|  | /* The rest of the dtv (including the generation counter) is | 
|  | Initialize with zero to indicate nothing there.  */ | 
|  |  | 
|  | /* Add the dtv to the thread data structures.  */ | 
|  | INSTALL_DTV (result, dtv); | 
|  | } | 
|  | else | 
|  | result = NULL; | 
|  |  | 
|  | return result; | 
|  | } | 
|  |  | 
|  |  | 
|  | /* Get size and alignment requirements of the static TLS block.  */ | 
|  | void | 
|  | internal_function | 
|  | _dl_get_tls_static_info (size_t *sizep, size_t *alignp) | 
|  | { | 
|  | *sizep = GL(dl_tls_static_size); | 
|  | *alignp = GL(dl_tls_static_align); | 
|  | } | 
|  |  | 
|  |  | 
|  | void * | 
|  | internal_function | 
|  | _dl_allocate_tls_storage (void) | 
|  | { | 
|  | void *result; | 
|  | size_t size = GL(dl_tls_static_size); | 
|  |  | 
|  | #if TLS_DTV_AT_TP | 
|  | /* Memory layout is: | 
|  | [ TLS_PRE_TCB_SIZE ] [ TLS_TCB_SIZE ] [ TLS blocks ] | 
|  | ^ This should be returned.  */ | 
|  | size += (TLS_PRE_TCB_SIZE + GL(dl_tls_static_align) - 1) | 
|  | & ~(GL(dl_tls_static_align) - 1); | 
|  | #endif | 
|  |  | 
|  | /* Allocate a correctly aligned chunk of memory.  */ | 
|  | result = __libc_memalign (GL(dl_tls_static_align), size); | 
|  | if (__builtin_expect (result != NULL, 1)) | 
|  | { | 
|  | /* Allocate the DTV.  */ | 
|  | void *allocated = result; | 
|  |  | 
|  | #if TLS_TCB_AT_TP | 
|  | /* The TCB follows the TLS blocks.  */ | 
|  | result = (char *) result + size - TLS_TCB_SIZE; | 
|  |  | 
|  | /* Clear the TCB data structure.  We can't ask the caller (i.e. | 
|  | libpthread) to do it, because we will initialize the DTV et al.  */ | 
|  | memset (result, '\0', TLS_TCB_SIZE); | 
|  | #elif TLS_DTV_AT_TP | 
|  | result = (char *) result + size - GL(dl_tls_static_size); | 
|  |  | 
|  | /* Clear the TCB data structure and TLS_PRE_TCB_SIZE bytes before it. | 
|  | We can't ask the caller (i.e. libpthread) to do it, because we will | 
|  | initialize the DTV et al.  */ | 
|  | memset ((char *) result - TLS_PRE_TCB_SIZE, '\0', | 
|  | TLS_PRE_TCB_SIZE + TLS_TCB_SIZE); | 
|  | #endif | 
|  |  | 
|  | result = allocate_dtv (result); | 
|  | if (result == NULL) | 
|  | free (allocated); | 
|  | } | 
|  |  | 
|  | return result; | 
|  | } | 
|  |  | 
|  |  | 
|  | #ifndef SHARED | 
|  | extern dtv_t _dl_static_dtv[]; | 
|  | # define _dl_initial_dtv (&_dl_static_dtv[1]) | 
|  | #endif | 
|  |  | 
|  | static dtv_t * | 
|  | _dl_resize_dtv (dtv_t *dtv) | 
|  | { | 
|  | /* Resize the dtv.  */ | 
|  | dtv_t *newp; | 
|  | /* Load GL(dl_tls_max_dtv_idx) atomically since it may be written to by | 
|  | other threads concurrently.  */ | 
|  | size_t newsize | 
|  | = atomic_load_acquire (&GL(dl_tls_max_dtv_idx)) + DTV_SURPLUS; | 
|  | size_t oldsize = dtv[-1].counter; | 
|  |  | 
|  | if (dtv == GL(dl_initial_dtv)) | 
|  | { | 
|  | /* This is the initial dtv that was either statically allocated in | 
|  | __libc_setup_tls or allocated during rtld startup using the | 
|  | dl-minimal.c malloc instead of the real malloc.  We can't free | 
|  | it, we have to abandon the old storage.  */ | 
|  |  | 
|  | newp = malloc ((2 + newsize) * sizeof (dtv_t)); | 
|  | if (newp == NULL) | 
|  | oom (); | 
|  | memcpy (newp, &dtv[-1], (2 + oldsize) * sizeof (dtv_t)); | 
|  | } | 
|  | else | 
|  | { | 
|  | newp = realloc (&dtv[-1], | 
|  | (2 + newsize) * sizeof (dtv_t)); | 
|  | if (newp == NULL) | 
|  | oom (); | 
|  | } | 
|  |  | 
|  | newp[0].counter = newsize; | 
|  |  | 
|  | /* Clear the newly allocated part.  */ | 
|  | memset (newp + 2 + oldsize, '\0', | 
|  | (newsize - oldsize) * sizeof (dtv_t)); | 
|  |  | 
|  | /* Return the generation counter.  */ | 
|  | return &newp[1]; | 
|  | } | 
|  |  | 
|  |  | 
|  | void * | 
|  | internal_function | 
|  | _dl_allocate_tls_init (void *result) | 
|  | { | 
|  | if (result == NULL) | 
|  | /* The memory allocation failed.  */ | 
|  | return NULL; | 
|  |  | 
|  | dtv_t *dtv = GET_DTV (result); | 
|  | struct dtv_slotinfo_list *listp; | 
|  | size_t total = 0; | 
|  | size_t maxgen = 0; | 
|  |  | 
|  | /* Check if the current dtv is big enough.   */ | 
|  | if (dtv[-1].counter < GL(dl_tls_max_dtv_idx)) | 
|  | { | 
|  | /* Resize the dtv.  */ | 
|  | dtv = _dl_resize_dtv (dtv); | 
|  |  | 
|  | /* Install this new dtv in the thread data structures.  */ | 
|  | INSTALL_DTV (result, &dtv[-1]); | 
|  | } | 
|  |  | 
|  | /* We have to prepare the dtv for all currently loaded modules using | 
|  | TLS.  For those which are dynamically loaded we add the values | 
|  | indicating deferred allocation.  */ | 
|  | listp = GL(dl_tls_dtv_slotinfo_list); | 
|  | while (1) | 
|  | { | 
|  | size_t cnt; | 
|  |  | 
|  | for (cnt = total == 0 ? 1 : 0; cnt < listp->len; ++cnt) | 
|  | { | 
|  | struct link_map *map; | 
|  | void *dest; | 
|  |  | 
|  | /* Check for the total number of used slots.  */ | 
|  | if (total + cnt > GL(dl_tls_max_dtv_idx)) | 
|  | break; | 
|  |  | 
|  | map = listp->slotinfo[cnt].map; | 
|  | if (map == NULL) | 
|  | /* Unused entry.  */ | 
|  | continue; | 
|  |  | 
|  | /* Keep track of the maximum generation number.  This might | 
|  | not be the generation counter.  */ | 
|  | assert (listp->slotinfo[cnt].gen <= GL(dl_tls_generation)); | 
|  | maxgen = MAX (maxgen, listp->slotinfo[cnt].gen); | 
|  |  | 
|  | dtv[map->l_tls_modid].pointer.val = TLS_DTV_UNALLOCATED; | 
|  | dtv[map->l_tls_modid].pointer.is_static = false; | 
|  |  | 
|  | if (map->l_tls_offset == NO_TLS_OFFSET | 
|  | || map->l_tls_offset == FORCED_DYNAMIC_TLS_OFFSET) | 
|  | continue; | 
|  |  | 
|  | assert (map->l_tls_modid == total + cnt); | 
|  | assert (map->l_tls_blocksize >= map->l_tls_initimage_size); | 
|  | #if TLS_TCB_AT_TP | 
|  | assert ((size_t) map->l_tls_offset >= map->l_tls_blocksize); | 
|  | dest = (char *) result - map->l_tls_offset; | 
|  | #elif TLS_DTV_AT_TP | 
|  | dest = (char *) result + map->l_tls_offset; | 
|  | #else | 
|  | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
|  | #endif | 
|  |  | 
|  | /* Copy the initialization image and clear the BSS part.  */ | 
|  | memset (__mempcpy (dest, map->l_tls_initimage, | 
|  | map->l_tls_initimage_size), '\0', | 
|  | map->l_tls_blocksize - map->l_tls_initimage_size); | 
|  | } | 
|  |  | 
|  | total += cnt; | 
|  | if (total >= GL(dl_tls_max_dtv_idx)) | 
|  | break; | 
|  |  | 
|  | listp = listp->next; | 
|  | assert (listp != NULL); | 
|  | } | 
|  |  | 
|  | /* The DTV version is up-to-date now.  */ | 
|  | dtv[0].counter = maxgen; | 
|  |  | 
|  | return result; | 
|  | } | 
|  | rtld_hidden_def (_dl_allocate_tls_init) | 
|  |  | 
|  | void * | 
|  | internal_function | 
|  | _dl_allocate_tls (void *mem) | 
|  | { | 
|  | return _dl_allocate_tls_init (mem == NULL | 
|  | ? _dl_allocate_tls_storage () | 
|  | : allocate_dtv (mem)); | 
|  | } | 
|  | rtld_hidden_def (_dl_allocate_tls) | 
|  |  | 
|  |  | 
|  | void | 
|  | internal_function | 
|  | _dl_deallocate_tls (void *tcb, bool dealloc_tcb) | 
|  | { | 
|  | dtv_t *dtv = GET_DTV (tcb); | 
|  |  | 
|  | /* We need to free the memory allocated for non-static TLS.  */ | 
|  | for (size_t cnt = 0; cnt < dtv[-1].counter; ++cnt) | 
|  | if (! dtv[1 + cnt].pointer.is_static | 
|  | && dtv[1 + cnt].pointer.val != TLS_DTV_UNALLOCATED) | 
|  | free (dtv[1 + cnt].pointer.val); | 
|  |  | 
|  | /* The array starts with dtv[-1].  */ | 
|  | if (dtv != GL(dl_initial_dtv)) | 
|  | free (dtv - 1); | 
|  |  | 
|  | if (dealloc_tcb) | 
|  | { | 
|  | #if TLS_TCB_AT_TP | 
|  | /* The TCB follows the TLS blocks.  Back up to free the whole block.  */ | 
|  | tcb -= GL(dl_tls_static_size) - TLS_TCB_SIZE; | 
|  | #elif TLS_DTV_AT_TP | 
|  | /* Back up the TLS_PRE_TCB_SIZE bytes.  */ | 
|  | tcb -= (TLS_PRE_TCB_SIZE + GL(dl_tls_static_align) - 1) | 
|  | & ~(GL(dl_tls_static_align) - 1); | 
|  | #endif | 
|  | free (tcb); | 
|  | } | 
|  | } | 
|  | rtld_hidden_def (_dl_deallocate_tls) | 
|  |  | 
|  |  | 
|  | #ifdef SHARED | 
|  | /* The __tls_get_addr function has two basic forms which differ in the | 
|  | arguments.  The IA-64 form takes two parameters, the module ID and | 
|  | offset.  The form used, among others, on IA-32 takes a reference to | 
|  | a special structure which contain the same information.  The second | 
|  | form seems to be more often used (in the moment) so we default to | 
|  | it.  Users of the IA-64 form have to provide adequate definitions | 
|  | of the following macros.  */ | 
|  | # ifndef GET_ADDR_ARGS | 
|  | #  define GET_ADDR_ARGS tls_index *ti | 
|  | #  define GET_ADDR_PARAM ti | 
|  | # endif | 
|  | # ifndef GET_ADDR_MODULE | 
|  | #  define GET_ADDR_MODULE ti->ti_module | 
|  | # endif | 
|  | # ifndef GET_ADDR_OFFSET | 
|  | #  define GET_ADDR_OFFSET ti->ti_offset | 
|  | # endif | 
|  |  | 
|  |  | 
|  | static void * | 
|  | allocate_and_init (struct link_map *map) | 
|  | { | 
|  | void *newp; | 
|  |  | 
|  | newp = __libc_memalign (map->l_tls_align, map->l_tls_blocksize); | 
|  | if (newp == NULL) | 
|  | oom (); | 
|  |  | 
|  | /* Initialize the memory.  */ | 
|  | memset (__mempcpy (newp, map->l_tls_initimage, map->l_tls_initimage_size), | 
|  | '\0', map->l_tls_blocksize - map->l_tls_initimage_size); | 
|  |  | 
|  | return newp; | 
|  | } | 
|  |  | 
|  |  | 
|  | struct link_map * | 
|  | _dl_update_slotinfo (unsigned long int req_modid) | 
|  | { | 
|  | struct link_map *the_map = NULL; | 
|  | dtv_t *dtv = THREAD_DTV (); | 
|  |  | 
|  | /* The global dl_tls_dtv_slotinfo array contains for each module | 
|  | index the generation counter current when the entry was created. | 
|  | This array never shrinks so that all module indices which were | 
|  | valid at some time can be used to access it.  Before the first | 
|  | use of a new module index in this function the array was extended | 
|  | appropriately.  Access also does not have to be guarded against | 
|  | modifications of the array.  It is assumed that pointer-size | 
|  | values can be read atomically even in SMP environments.  It is | 
|  | possible that other threads at the same time dynamically load | 
|  | code and therefore add to the slotinfo list.  This is a problem | 
|  | since we must not pick up any information about incomplete work. | 
|  | The solution to this is to ignore all dtv slots which were | 
|  | created after the one we are currently interested.  We know that | 
|  | dynamic loading for this module is completed and this is the last | 
|  | load operation we know finished.  */ | 
|  | unsigned long int idx = req_modid; | 
|  | struct dtv_slotinfo_list *listp = GL(dl_tls_dtv_slotinfo_list); | 
|  |  | 
|  | while (idx >= listp->len) | 
|  | { | 
|  | idx -= listp->len; | 
|  | listp = listp->next; | 
|  | } | 
|  |  | 
|  | if (dtv[0].counter < listp->slotinfo[idx].gen) | 
|  | { | 
|  | /* The generation counter for the slot is higher than what the | 
|  | current dtv implements.  We have to update the whole dtv but | 
|  | only those entries with a generation counter <= the one for | 
|  | the entry we need.  */ | 
|  | size_t new_gen = listp->slotinfo[idx].gen; | 
|  | size_t total = 0; | 
|  |  | 
|  | /* We have to look through the entire dtv slotinfo list.  */ | 
|  | listp =  GL(dl_tls_dtv_slotinfo_list); | 
|  | do | 
|  | { | 
|  | for (size_t cnt = total == 0 ? 1 : 0; cnt < listp->len; ++cnt) | 
|  | { | 
|  | size_t gen = listp->slotinfo[cnt].gen; | 
|  |  | 
|  | if (gen > new_gen) | 
|  | /* This is a slot for a generation younger than the | 
|  | one we are handling now.  It might be incompletely | 
|  | set up so ignore it.  */ | 
|  | continue; | 
|  |  | 
|  | /* If the entry is older than the current dtv layout we | 
|  | know we don't have to handle it.  */ | 
|  | if (gen <= dtv[0].counter) | 
|  | continue; | 
|  |  | 
|  | /* If there is no map this means the entry is empty.  */ | 
|  | struct link_map *map = listp->slotinfo[cnt].map; | 
|  | if (map == NULL) | 
|  | { | 
|  | if (dtv[-1].counter >= total + cnt) | 
|  | { | 
|  | /* If this modid was used at some point the memory | 
|  | might still be allocated.  */ | 
|  | if (! dtv[total + cnt].pointer.is_static | 
|  | && (dtv[total + cnt].pointer.val | 
|  | != TLS_DTV_UNALLOCATED)) | 
|  | free (dtv[total + cnt].pointer.val); | 
|  | dtv[total + cnt].pointer.val = TLS_DTV_UNALLOCATED; | 
|  | dtv[total + cnt].pointer.is_static = false; | 
|  | } | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* Check whether the current dtv array is large enough.  */ | 
|  | size_t modid = map->l_tls_modid; | 
|  | assert (total + cnt == modid); | 
|  | if (dtv[-1].counter < modid) | 
|  | { | 
|  | /* Resize the dtv.  */ | 
|  | dtv = _dl_resize_dtv (dtv); | 
|  |  | 
|  | assert (modid <= dtv[-1].counter); | 
|  |  | 
|  | /* Install this new dtv in the thread data | 
|  | structures.  */ | 
|  | INSTALL_NEW_DTV (dtv); | 
|  | } | 
|  |  | 
|  | /* If there is currently memory allocate for this | 
|  | dtv entry free it.  */ | 
|  | /* XXX Ideally we will at some point create a memory | 
|  | pool.  */ | 
|  | if (! dtv[modid].pointer.is_static | 
|  | && dtv[modid].pointer.val != TLS_DTV_UNALLOCATED) | 
|  | /* Note that free is called for NULL is well.  We | 
|  | deallocate even if it is this dtv entry we are | 
|  | supposed to load.  The reason is that we call | 
|  | memalign and not malloc.  */ | 
|  | free (dtv[modid].pointer.val); | 
|  |  | 
|  | dtv[modid].pointer.val = TLS_DTV_UNALLOCATED; | 
|  | dtv[modid].pointer.is_static = false; | 
|  |  | 
|  | if (modid == req_modid) | 
|  | the_map = map; | 
|  | } | 
|  |  | 
|  | total += listp->len; | 
|  | } | 
|  | while ((listp = listp->next) != NULL); | 
|  |  | 
|  | /* This will be the new maximum generation counter.  */ | 
|  | dtv[0].counter = new_gen; | 
|  | } | 
|  |  | 
|  | return the_map; | 
|  | } | 
|  |  | 
|  |  | 
|  | static void * | 
|  | __attribute_noinline__ | 
|  | tls_get_addr_tail (GET_ADDR_ARGS, dtv_t *dtv, struct link_map *the_map) | 
|  | { | 
|  | /* The allocation was deferred.  Do it now.  */ | 
|  | if (the_map == NULL) | 
|  | { | 
|  | /* Find the link map for this module.  */ | 
|  | size_t idx = GET_ADDR_MODULE; | 
|  | struct dtv_slotinfo_list *listp = GL(dl_tls_dtv_slotinfo_list); | 
|  |  | 
|  | while (idx >= listp->len) | 
|  | { | 
|  | idx -= listp->len; | 
|  | listp = listp->next; | 
|  | } | 
|  |  | 
|  | the_map = listp->slotinfo[idx].map; | 
|  | } | 
|  |  | 
|  | /* Make sure that, if a dlopen running in parallel forces the | 
|  | variable into static storage, we'll wait until the address in the | 
|  | static TLS block is set up, and use that.  If we're undecided | 
|  | yet, make sure we make the decision holding the lock as well.  */ | 
|  | if (__glibc_unlikely (the_map->l_tls_offset | 
|  | != FORCED_DYNAMIC_TLS_OFFSET)) | 
|  | { | 
|  | __rtld_lock_lock_recursive (GL(dl_load_lock)); | 
|  | if (__glibc_likely (the_map->l_tls_offset == NO_TLS_OFFSET)) | 
|  | { | 
|  | the_map->l_tls_offset = FORCED_DYNAMIC_TLS_OFFSET; | 
|  | __rtld_lock_unlock_recursive (GL(dl_load_lock)); | 
|  | } | 
|  | else if (__glibc_likely (the_map->l_tls_offset | 
|  | != FORCED_DYNAMIC_TLS_OFFSET)) | 
|  | { | 
|  | #if TLS_TCB_AT_TP | 
|  | void *p = (char *) THREAD_SELF - the_map->l_tls_offset; | 
|  | #elif TLS_DTV_AT_TP | 
|  | void *p = (char *) THREAD_SELF + the_map->l_tls_offset + TLS_PRE_TCB_SIZE; | 
|  | #else | 
|  | # error "Either TLS_TCB_AT_TP or TLS_DTV_AT_TP must be defined" | 
|  | #endif | 
|  | __rtld_lock_unlock_recursive (GL(dl_load_lock)); | 
|  |  | 
|  | dtv[GET_ADDR_MODULE].pointer.is_static = true; | 
|  | dtv[GET_ADDR_MODULE].pointer.val = p; | 
|  |  | 
|  | return (char *) p + GET_ADDR_OFFSET; | 
|  | } | 
|  | else | 
|  | __rtld_lock_unlock_recursive (GL(dl_load_lock)); | 
|  | } | 
|  | void *p = dtv[GET_ADDR_MODULE].pointer.val = allocate_and_init (the_map); | 
|  | assert (!dtv[GET_ADDR_MODULE].pointer.is_static); | 
|  |  | 
|  | return (char *) p + GET_ADDR_OFFSET; | 
|  | } | 
|  |  | 
|  |  | 
|  | static struct link_map * | 
|  | __attribute_noinline__ | 
|  | update_get_addr (GET_ADDR_ARGS) | 
|  | { | 
|  | struct link_map *the_map = _dl_update_slotinfo (GET_ADDR_MODULE); | 
|  | dtv_t *dtv = THREAD_DTV (); | 
|  |  | 
|  | void *p = dtv[GET_ADDR_MODULE].pointer.val; | 
|  |  | 
|  | if (__glibc_unlikely (p == TLS_DTV_UNALLOCATED)) | 
|  | return tls_get_addr_tail (GET_ADDR_PARAM, dtv, the_map); | 
|  |  | 
|  | return (void *) p + GET_ADDR_OFFSET; | 
|  | } | 
|  |  | 
|  | /* For all machines that have a non-macro version of __tls_get_addr, we | 
|  | want to use rtld_hidden_proto/rtld_hidden_def in order to call the | 
|  | internal alias for __tls_get_addr from ld.so. This avoids a PLT entry | 
|  | in ld.so for __tls_get_addr.  */ | 
|  |  | 
|  | #ifndef __tls_get_addr | 
|  | extern void * __tls_get_addr (GET_ADDR_ARGS); | 
|  | rtld_hidden_proto (__tls_get_addr) | 
|  | rtld_hidden_def (__tls_get_addr) | 
|  | #endif | 
|  |  | 
|  | /* The generic dynamic and local dynamic model cannot be used in | 
|  | statically linked applications.  */ | 
|  | void * | 
|  | __tls_get_addr (GET_ADDR_ARGS) | 
|  | { | 
|  | dtv_t *dtv = THREAD_DTV (); | 
|  |  | 
|  | if (__glibc_unlikely (dtv[0].counter != GL(dl_tls_generation))) | 
|  | return update_get_addr (GET_ADDR_PARAM); | 
|  |  | 
|  | void *p = dtv[GET_ADDR_MODULE].pointer.val; | 
|  |  | 
|  | if (__glibc_unlikely (p == TLS_DTV_UNALLOCATED)) | 
|  | return tls_get_addr_tail (GET_ADDR_PARAM, dtv, NULL); | 
|  |  | 
|  | return (char *) p + GET_ADDR_OFFSET; | 
|  | } | 
|  | #endif | 
|  |  | 
|  |  | 
|  | /* Look up the module's TLS block as for __tls_get_addr, | 
|  | but never touch anything.  Return null if it's not allocated yet.  */ | 
|  | void * | 
|  | _dl_tls_get_addr_soft (struct link_map *l) | 
|  | { | 
|  | if (__glibc_unlikely (l->l_tls_modid == 0)) | 
|  | /* This module has no TLS segment.  */ | 
|  | return NULL; | 
|  |  | 
|  | dtv_t *dtv = THREAD_DTV (); | 
|  | if (__glibc_unlikely (dtv[0].counter != GL(dl_tls_generation))) | 
|  | { | 
|  | /* This thread's DTV is not completely current, | 
|  | but it might already cover this module.  */ | 
|  |  | 
|  | if (l->l_tls_modid >= dtv[-1].counter) | 
|  | /* Nope.  */ | 
|  | return NULL; | 
|  |  | 
|  | size_t idx = l->l_tls_modid; | 
|  | struct dtv_slotinfo_list *listp = GL(dl_tls_dtv_slotinfo_list); | 
|  | while (idx >= listp->len) | 
|  | { | 
|  | idx -= listp->len; | 
|  | listp = listp->next; | 
|  | } | 
|  |  | 
|  | /* We've reached the slot for this module. | 
|  | If its generation counter is higher than the DTV's, | 
|  | this thread does not know about this module yet.  */ | 
|  | if (dtv[0].counter < listp->slotinfo[idx].gen) | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | void *data = dtv[l->l_tls_modid].pointer.val; | 
|  | if (__glibc_unlikely (data == TLS_DTV_UNALLOCATED)) | 
|  | /* The DTV is current, but this thread has not yet needed | 
|  | to allocate this module's segment.  */ | 
|  | data = NULL; | 
|  |  | 
|  | return data; | 
|  | } | 
|  |  | 
|  |  | 
|  | void | 
|  | _dl_add_to_slotinfo (struct link_map *l) | 
|  | { | 
|  | /* Now that we know the object is loaded successfully add | 
|  | modules containing TLS data to the dtv info table.  We | 
|  | might have to increase its size.  */ | 
|  | struct dtv_slotinfo_list *listp; | 
|  | struct dtv_slotinfo_list *prevp; | 
|  | size_t idx = l->l_tls_modid; | 
|  |  | 
|  | /* Find the place in the dtv slotinfo list.  */ | 
|  | listp = GL(dl_tls_dtv_slotinfo_list); | 
|  | prevp = NULL;		/* Needed to shut up gcc.  */ | 
|  | do | 
|  | { | 
|  | /* Does it fit in the array of this list element?  */ | 
|  | if (idx < listp->len) | 
|  | break; | 
|  | idx -= listp->len; | 
|  | prevp = listp; | 
|  | listp = listp->next; | 
|  | } | 
|  | while (listp != NULL); | 
|  |  | 
|  | if (listp == NULL) | 
|  | { | 
|  | /* When we come here it means we have to add a new element | 
|  | to the slotinfo list.  And the new module must be in | 
|  | the first slot.  */ | 
|  | assert (idx == 0); | 
|  |  | 
|  | listp = prevp->next = (struct dtv_slotinfo_list *) | 
|  | malloc (sizeof (struct dtv_slotinfo_list) | 
|  | + TLS_SLOTINFO_SURPLUS * sizeof (struct dtv_slotinfo)); | 
|  | if (listp == NULL) | 
|  | { | 
|  | /* We ran out of memory.  We will simply fail this | 
|  | call but don't undo anything we did so far.  The | 
|  | application will crash or be terminated anyway very | 
|  | soon.  */ | 
|  |  | 
|  | /* We have to do this since some entries in the dtv | 
|  | slotinfo array might already point to this | 
|  | generation.  */ | 
|  | ++GL(dl_tls_generation); | 
|  |  | 
|  | _dl_signal_error (ENOMEM, "dlopen", NULL, N_("\ | 
|  | cannot create TLS data structures")); | 
|  | } | 
|  |  | 
|  | listp->len = TLS_SLOTINFO_SURPLUS; | 
|  | listp->next = NULL; | 
|  | memset (listp->slotinfo, '\0', | 
|  | TLS_SLOTINFO_SURPLUS * sizeof (struct dtv_slotinfo)); | 
|  | } | 
|  |  | 
|  | /* Add the information into the slotinfo data structure.  */ | 
|  | listp->slotinfo[idx].map = l; | 
|  | listp->slotinfo[idx].gen = GL(dl_tls_generation) + 1; | 
|  | } |