Home | History | Annotate | Download | only in linker
      1 /*
      2  * Copyright (C) 2008, 2009 The Android Open Source Project
      3  * All rights reserved.
      4  *
      5  * Redistribution and use in source and binary forms, with or without
      6  * modification, are permitted provided that the following conditions
      7  * are met:
      8  *  * Redistributions of source code must retain the above copyright
      9  *    notice, this list of conditions and the following disclaimer.
     10  *  * Redistributions in binary form must reproduce the above copyright
     11  *    notice, this list of conditions and the following disclaimer in
     12  *    the documentation and/or other materials provided with the
     13  *    distribution.
     14  *
     15  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     16  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     17  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
     18  * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
     19  * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
     20  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
     21  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
     22  * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
     23  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
     24  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
     25  * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     26  * SUCH DAMAGE.
     27  */
     28 
     29 #include <linux/auxvec.h>
     30 
     31 #include <stdio.h>
     32 #include <stdlib.h>
     33 #include <string.h>
     34 #include <unistd.h>
     35 #include <fcntl.h>
     36 #include <errno.h>
     37 #include <dlfcn.h>
     38 #include <sys/stat.h>
     39 
     40 #include <pthread.h>
     41 
     42 #include <sys/mman.h>
     43 
     44 #include <sys/atomics.h>
     45 
     46 /* special private C library header - see Android.mk */
     47 #include <bionic_tls.h>
     48 
     49 #include "linker.h"
     50 #include "linker_debug.h"
     51 #include "linker_format.h"
     52 
     53 #include "ba.h"
     54 
     55 #define ALLOW_SYMBOLS_FROM_MAIN 1
     56 #define SO_MAX 96
     57 
     58 /* Assume average path length of 64 and max 8 paths */
     59 #define LDPATH_BUFSIZE 512
     60 #define LDPATH_MAX 8
     61 
     62 #define LDPRELOAD_BUFSIZE 512
     63 #define LDPRELOAD_MAX 8
     64 
     65 /* >>> IMPORTANT NOTE - READ ME BEFORE MODIFYING <<<
     66  *
     67  * Do NOT use malloc() and friends or pthread_*() code here.
     68  * Don't use printf() either; it's caused mysterious memory
     69  * corruption in the past.
     70  * The linker runs before we bring up libc and it's easiest
     71  * to make sure it does not depend on any complex libc features
     72  *
     73  * open issues / todo:
     74  *
     75  * - are we doing everything we should for ARM_COPY relocations?
     76  * - cleaner error reporting
     77  * - after linking, set as much stuff as possible to READONLY
     78  *   and NOEXEC
     79  * - linker hardcodes PAGE_SIZE and PAGE_MASK because the kernel
     80  *   headers provide versions that are negative...
     81  * - allocate space for soinfo structs dynamically instead of
     82  *   having a hard limit (64)
     83 */
     84 
     85 
     86 static int link_image(soinfo *si, unsigned wr_offset);
     87 
     88 static int socount = 0;
     89 static soinfo sopool[SO_MAX];
     90 static soinfo *freelist = NULL;
     91 static soinfo *solist = &libdl_info;
     92 static soinfo *sonext = &libdl_info;
     93 #if ALLOW_SYMBOLS_FROM_MAIN
     94 static soinfo *somain; /* main process, always the one after libdl_info */
     95 #endif
     96 
     97 
     98 /* Set up for the buddy allocator managing the non-prelinked libraries. */
     99 static struct ba_bits ba_nonprelink_bitmap[(LIBLAST - LIBBASE) / LIBINC];
    100 static struct ba ba_nonprelink = {
    101     .base = LIBBASE,
    102     .size = LIBLAST - LIBBASE,
    103     .min_alloc = LIBINC,
    104     /* max_order will be determined automatically */
    105     .bitmap = ba_nonprelink_bitmap,
    106     .num_entries = sizeof(ba_nonprelink_bitmap)/sizeof(ba_nonprelink_bitmap[0]),
    107 };
    108 
    109 static inline int validate_soinfo(soinfo *si)
    110 {
    111     return (si >= sopool && si < sopool + SO_MAX) ||
    112         si == &libdl_info;
    113 }
    114 
    115 static char ldpaths_buf[LDPATH_BUFSIZE];
    116 static const char *ldpaths[LDPATH_MAX + 1];
    117 
    118 static char ldpreloads_buf[LDPRELOAD_BUFSIZE];
    119 static const char *ldpreload_names[LDPRELOAD_MAX + 1];
    120 
    121 static soinfo *preloads[LDPRELOAD_MAX + 1];
    122 
    123 int debug_verbosity;
    124 static int pid;
    125 
    126 #if STATS
    127 struct _link_stats linker_stats;
    128 #endif
    129 
    130 #if COUNT_PAGES
    131 unsigned bitmask[4096];
    132 #endif
    133 
    134 #ifndef PT_ARM_EXIDX
    135 #define PT_ARM_EXIDX    0x70000001      /* .ARM.exidx segment */
    136 #endif
    137 
    138 #define HOODLUM(name, ret, ...)                                               \
    139     ret name __VA_ARGS__                                                      \
    140     {                                                                         \
    141         char errstr[] = "ERROR: " #name " called from the dynamic linker!\n"; \
    142         write(2, errstr, sizeof(errstr));                                     \
    143         abort();                                                              \
    144     }
    145 HOODLUM(malloc, void *, (size_t size));
    146 HOODLUM(free, void, (void *ptr));
    147 HOODLUM(realloc, void *, (void *ptr, size_t size));
    148 HOODLUM(calloc, void *, (size_t cnt, size_t size));
    149 
    150 static char tmp_err_buf[768];
    151 static char __linker_dl_err_buf[768];
    152 #define DL_ERR(fmt, x...)                                                     \
    153     do {                                                                      \
    154         format_buffer(__linker_dl_err_buf, sizeof(__linker_dl_err_buf),            \
    155                  "%s[%d]: " fmt, __func__, __LINE__, ##x);                    \
    156         ERROR(fmt "\n", ##x);                                                      \
    157     } while(0)
    158 
    159 const char *linker_get_error(void)
    160 {
    161     return (const char *)&__linker_dl_err_buf[0];
    162 }
    163 
    164 /*
    165  * This function is an empty stub where GDB locates a breakpoint to get notified
    166  * about linker activity.
    167  */
    168 extern void __attribute__((noinline)) rtld_db_dlactivity(void);
    169 
    170 static struct r_debug _r_debug = {1, NULL, &rtld_db_dlactivity,
    171                                   RT_CONSISTENT, 0};
    172 static struct link_map *r_debug_tail = 0;
    173 
    174 static pthread_mutex_t _r_debug_lock = PTHREAD_MUTEX_INITIALIZER;
    175 
    176 static void insert_soinfo_into_debug_map(soinfo * info)
    177 {
    178     struct link_map * map;
    179 
    180     /* Copy the necessary fields into the debug structure.
    181      */
    182     map = &(info->linkmap);
    183     map->l_addr = info->base;
    184     map->l_name = (char*) info->name;
    185     map->l_ld = (uintptr_t)info->dynamic;
    186 
    187     /* Stick the new library at the end of the list.
    188      * gdb tends to care more about libc than it does
    189      * about leaf libraries, and ordering it this way
    190      * reduces the back-and-forth over the wire.
    191      */
    192     if (r_debug_tail) {
    193         r_debug_tail->l_next = map;
    194         map->l_prev = r_debug_tail;
    195         map->l_next = 0;
    196     } else {
    197         _r_debug.r_map = map;
    198         map->l_prev = 0;
    199         map->l_next = 0;
    200     }
    201     r_debug_tail = map;
    202 }
    203 
    204 static void remove_soinfo_from_debug_map(soinfo * info)
    205 {
    206     struct link_map * map = &(info->linkmap);
    207 
    208     if (r_debug_tail == map)
    209         r_debug_tail = map->l_prev;
    210 
    211     if (map->l_prev) map->l_prev->l_next = map->l_next;
    212     if (map->l_next) map->l_next->l_prev = map->l_prev;
    213 }
    214 
    215 void notify_gdb_of_load(soinfo * info)
    216 {
    217     if (info->flags & FLAG_EXE) {
    218         // GDB already knows about the main executable
    219         return;
    220     }
    221 
    222     pthread_mutex_lock(&_r_debug_lock);
    223 
    224     _r_debug.r_state = RT_ADD;
    225     rtld_db_dlactivity();
    226 
    227     insert_soinfo_into_debug_map(info);
    228 
    229     _r_debug.r_state = RT_CONSISTENT;
    230     rtld_db_dlactivity();
    231 
    232     pthread_mutex_unlock(&_r_debug_lock);
    233 }
    234 
    235 void notify_gdb_of_unload(soinfo * info)
    236 {
    237     if (info->flags & FLAG_EXE) {
    238         // GDB already knows about the main executable
    239         return;
    240     }
    241 
    242     pthread_mutex_lock(&_r_debug_lock);
    243 
    244     _r_debug.r_state = RT_DELETE;
    245     rtld_db_dlactivity();
    246 
    247     remove_soinfo_from_debug_map(info);
    248 
    249     _r_debug.r_state = RT_CONSISTENT;
    250     rtld_db_dlactivity();
    251 
    252     pthread_mutex_unlock(&_r_debug_lock);
    253 }
    254 
    255 void notify_gdb_of_libraries()
    256 {
    257     _r_debug.r_state = RT_ADD;
    258     rtld_db_dlactivity();
    259     _r_debug.r_state = RT_CONSISTENT;
    260     rtld_db_dlactivity();
    261 }
    262 
    263 static soinfo *alloc_info(const char *name)
    264 {
    265     soinfo *si;
    266 
    267     if(strlen(name) >= SOINFO_NAME_LEN) {
    268         DL_ERR("%5d library name %s too long", pid, name);
    269         return NULL;
    270     }
    271 
    272     /* The freelist is populated when we call free_info(), which in turn is
    273        done only by dlclose(), which is not likely to be used.
    274     */
    275     if (!freelist) {
    276         if(socount == SO_MAX) {
    277             DL_ERR("%5d too many libraries when loading %s", pid, name);
    278             return NULL;
    279         }
    280         freelist = sopool + socount++;
    281         freelist->next = NULL;
    282     }
    283 
    284     si = freelist;
    285     freelist = freelist->next;
    286 
    287     /* Make sure we get a clean block of soinfo */
    288     memset(si, 0, sizeof(soinfo));
    289     strcpy((char*) si->name, name);
    290     sonext->next = si;
    291     si->ba_index = -1; /* by default, prelinked */
    292     si->next = NULL;
    293     si->refcount = 0;
    294     sonext = si;
    295 
    296     TRACE("%5d name %s: allocated soinfo @ %p\n", pid, name, si);
    297     return si;
    298 }
    299 
    300 static void free_info(soinfo *si)
    301 {
    302     soinfo *prev = NULL, *trav;
    303 
    304     TRACE("%5d name %s: freeing soinfo @ %p\n", pid, si->name, si);
    305 
    306     for(trav = solist; trav != NULL; trav = trav->next){
    307         if (trav == si)
    308             break;
    309         prev = trav;
    310     }
    311     if (trav == NULL) {
    312         /* si was not ni solist */
    313         DL_ERR("%5d name %s is not in solist!", pid, si->name);
    314         return;
    315     }
    316 
    317     /* prev will never be NULL, because the first entry in solist is
    318        always the static libdl_info.
    319     */
    320     prev->next = si->next;
    321     if (si == sonext) sonext = prev;
    322     si->next = freelist;
    323     freelist = si;
    324 }
    325 
    326 #ifndef LINKER_TEXT_BASE
    327 #error "linker's makefile must define LINKER_TEXT_BASE"
    328 #endif
    329 #ifndef LINKER_AREA_SIZE
    330 #error "linker's makefile must define LINKER_AREA_SIZE"
    331 #endif
    332 #define LINKER_BASE ((LINKER_TEXT_BASE) & 0xfff00000)
    333 #define LINKER_TOP  (LINKER_BASE + (LINKER_AREA_SIZE))
    334 
    335 const char *addr_to_name(unsigned addr)
    336 {
    337     soinfo *si;
    338 
    339     for(si = solist; si != 0; si = si->next){
    340         if((addr >= si->base) && (addr < (si->base + si->size))) {
    341             return si->name;
    342         }
    343     }
    344 
    345     if((addr >= LINKER_BASE) && (addr < LINKER_TOP)){
    346         return "linker";
    347     }
    348 
    349     return "";
    350 }
    351 
    352 /* For a given PC, find the .so that it belongs to.
    353  * Returns the base address of the .ARM.exidx section
    354  * for that .so, and the number of 8-byte entries
    355  * in that section (via *pcount).
    356  *
    357  * Intended to be called by libc's __gnu_Unwind_Find_exidx().
    358  *
    359  * This function is exposed via dlfcn.c and libdl.so.
    360  */
    361 #ifdef ANDROID_ARM_LINKER
    362 _Unwind_Ptr dl_unwind_find_exidx(_Unwind_Ptr pc, int *pcount)
    363 {
    364     soinfo *si;
    365     unsigned addr = (unsigned)pc;
    366 
    367     if ((addr < LINKER_BASE) || (addr >= LINKER_TOP)) {
    368         for (si = solist; si != 0; si = si->next){
    369             if ((addr >= si->base) && (addr < (si->base + si->size))) {
    370                 *pcount = si->ARM_exidx_count;
    371                 return (_Unwind_Ptr)(si->base + (unsigned long)si->ARM_exidx);
    372             }
    373         }
    374     }
    375    *pcount = 0;
    376     return NULL;
    377 }
    378 #elif defined(ANDROID_X86_LINKER) || defined(ANDROID_SH_LINKER)
    379 /* Here, we only have to provide a callback to iterate across all the
    380  * loaded libraries. gcc_eh does the rest. */
    381 int
    382 dl_iterate_phdr(int (*cb)(struct dl_phdr_info *info, size_t size, void *data),
    383                 void *data)
    384 {
    385     soinfo *si;
    386     struct dl_phdr_info dl_info;
    387     int rv = 0;
    388 
    389     for (si = solist; si != NULL; si = si->next) {
    390         dl_info.dlpi_addr = si->linkmap.l_addr;
    391         dl_info.dlpi_name = si->linkmap.l_name;
    392         dl_info.dlpi_phdr = si->phdr;
    393         dl_info.dlpi_phnum = si->phnum;
    394         rv = cb(&dl_info, sizeof (struct dl_phdr_info), data);
    395         if (rv != 0)
    396             break;
    397     }
    398     return rv;
    399 }
    400 #endif
    401 
    402 static Elf32_Sym *_elf_lookup(soinfo *si, unsigned hash, const char *name)
    403 {
    404     Elf32_Sym *s;
    405     Elf32_Sym *symtab = si->symtab;
    406     const char *strtab = si->strtab;
    407     unsigned n;
    408 
    409     TRACE_TYPE(LOOKUP, "%5d SEARCH %s in %s@0x%08x %08x %d\n", pid,
    410                name, si->name, si->base, hash, hash % si->nbucket);
    411     n = hash % si->nbucket;
    412 
    413     for(n = si->bucket[hash % si->nbucket]; n != 0; n = si->chain[n]){
    414         s = symtab + n;
    415         if(strcmp(strtab + s->st_name, name)) continue;
    416 
    417             /* only concern ourselves with global and weak symbol definitions */
    418         switch(ELF32_ST_BIND(s->st_info)){
    419         case STB_GLOBAL:
    420         case STB_WEAK:
    421                 /* no section == undefined */
    422             if(s->st_shndx == 0) continue;
    423 
    424             TRACE_TYPE(LOOKUP, "%5d FOUND %s in %s (%08x) %d\n", pid,
    425                        name, si->name, s->st_value, s->st_size);
    426             return s;
    427         }
    428     }
    429 
    430     return NULL;
    431 }
    432 
    433 static unsigned elfhash(const char *_name)
    434 {
    435     const unsigned char *name = (const unsigned char *) _name;
    436     unsigned h = 0, g;
    437 
    438     while(*name) {
    439         h = (h << 4) + *name++;
    440         g = h & 0xf0000000;
    441         h ^= g;
    442         h ^= g >> 24;
    443     }
    444     return h;
    445 }
    446 
    447 static Elf32_Sym *
    448 _do_lookup(soinfo *si, const char *name, unsigned *base)
    449 {
    450     unsigned elf_hash = elfhash(name);
    451     Elf32_Sym *s;
    452     unsigned *d;
    453     soinfo *lsi = si;
    454     int i;
    455 
    456     /* Look for symbols in the local scope first (the object who is
    457      * searching). This happens with C++ templates on i386 for some
    458      * reason.
    459      *
    460      * Notes on weak symbols:
    461      * The ELF specs are ambigious about treatment of weak definitions in
    462      * dynamic linking.  Some systems return the first definition found
    463      * and some the first non-weak definition.   This is system dependent.
    464      * Here we return the first definition found for simplicity.  */
    465     s = _elf_lookup(si, elf_hash, name);
    466     if(s != NULL)
    467         goto done;
    468 
    469     /* Next, look for it in the preloads list */
    470     for(i = 0; preloads[i] != NULL; i++) {
    471         lsi = preloads[i];
    472         s = _elf_lookup(lsi, elf_hash, name);
    473         if(s != NULL)
    474             goto done;
    475     }
    476 
    477     for(d = si->dynamic; *d; d += 2) {
    478         if(d[0] == DT_NEEDED){
    479             lsi = (soinfo *)d[1];
    480             if (!validate_soinfo(lsi)) {
    481                 DL_ERR("%5d bad DT_NEEDED pointer in %s",
    482                        pid, si->name);
    483                 return NULL;
    484             }
    485 
    486             DEBUG("%5d %s: looking up %s in %s\n",
    487                   pid, si->name, name, lsi->name);
    488             s = _elf_lookup(lsi, elf_hash, name);
    489             if ((s != NULL) && (s->st_shndx != SHN_UNDEF))
    490                 goto done;
    491         }
    492     }
    493 
    494 #if ALLOW_SYMBOLS_FROM_MAIN
    495     /* If we are resolving relocations while dlopen()ing a library, it's OK for
    496      * the library to resolve a symbol that's defined in the executable itself,
    497      * although this is rare and is generally a bad idea.
    498      */
    499     if (somain) {
    500         lsi = somain;
    501         DEBUG("%5d %s: looking up %s in executable %s\n",
    502               pid, si->name, name, lsi->name);
    503         s = _elf_lookup(lsi, elf_hash, name);
    504     }
    505 #endif
    506 
    507 done:
    508     if(s != NULL) {
    509         TRACE_TYPE(LOOKUP, "%5d si %s sym %s s->st_value = 0x%08x, "
    510                    "found in %s, base = 0x%08x\n",
    511                    pid, si->name, name, s->st_value, lsi->name, lsi->base);
    512         *base = lsi->base;
    513         return s;
    514     }
    515 
    516     return NULL;
    517 }
    518 
    519 /* This is used by dl_sym().  It performs symbol lookup only within the
    520    specified soinfo object and not in any of its dependencies.
    521  */
    522 Elf32_Sym *lookup_in_library(soinfo *si, const char *name)
    523 {
    524     return _elf_lookup(si, elfhash(name), name);
    525 }
    526 
    527 /* This is used by dl_sym().  It performs a global symbol lookup.
    528  */
    529 Elf32_Sym *lookup(const char *name, soinfo **found, soinfo *start)
    530 {
    531     unsigned elf_hash = elfhash(name);
    532     Elf32_Sym *s = NULL;
    533     soinfo *si;
    534 
    535     if(start == NULL) {
    536         start = solist;
    537     }
    538 
    539     for(si = start; (s == NULL) && (si != NULL); si = si->next)
    540     {
    541         if(si->flags & FLAG_ERROR)
    542             continue;
    543         s = _elf_lookup(si, elf_hash, name);
    544         if (s != NULL) {
    545             *found = si;
    546             break;
    547         }
    548     }
    549 
    550     if(s != NULL) {
    551         TRACE_TYPE(LOOKUP, "%5d %s s->st_value = 0x%08x, "
    552                    "si->base = 0x%08x\n", pid, name, s->st_value, si->base);
    553         return s;
    554     }
    555 
    556     return NULL;
    557 }
    558 
    559 soinfo *find_containing_library(void *addr)
    560 {
    561     soinfo *si;
    562 
    563     for(si = solist; si != NULL; si = si->next)
    564     {
    565         if((unsigned)addr >= si->base && (unsigned)addr - si->base < si->size) {
    566             return si;
    567         }
    568     }
    569 
    570     return NULL;
    571 }
    572 
    573 Elf32_Sym *find_containing_symbol(void *addr, soinfo *si)
    574 {
    575     unsigned int i;
    576     unsigned soaddr = (unsigned)addr - si->base;
    577 
    578     /* Search the library's symbol table for any defined symbol which
    579      * contains this address */
    580     for(i=0; i<si->nchain; i++) {
    581         Elf32_Sym *sym = &si->symtab[i];
    582 
    583         if(sym->st_shndx != SHN_UNDEF &&
    584            soaddr >= sym->st_value &&
    585            soaddr < sym->st_value + sym->st_size) {
    586             return sym;
    587         }
    588     }
    589 
    590     return NULL;
    591 }
    592 
    593 #if 0
    594 static void dump(soinfo *si)
    595 {
    596     Elf32_Sym *s = si->symtab;
    597     unsigned n;
    598 
    599     for(n = 0; n < si->nchain; n++) {
    600         TRACE("%5d %04d> %08x: %02x %04x %08x %08x %s\n", pid, n, s,
    601                s->st_info, s->st_shndx, s->st_value, s->st_size,
    602                si->strtab + s->st_name);
    603         s++;
    604     }
    605 }
    606 #endif
    607 
    608 static const char *sopaths[] = {
    609     "/vendor/lib",
    610     "/system/lib",
    611     0
    612 };
    613 
    614 static int _open_lib(const char *name)
    615 {
    616     int fd;
    617     struct stat filestat;
    618 
    619     if ((stat(name, &filestat) >= 0) && S_ISREG(filestat.st_mode)) {
    620         if ((fd = open(name, O_RDONLY)) >= 0)
    621             return fd;
    622     }
    623 
    624     return -1;
    625 }
    626 
    627 static int open_library(const char *name)
    628 {
    629     int fd;
    630     char buf[512];
    631     const char **path;
    632     int n;
    633 
    634     TRACE("[ %5d opening %s ]\n", pid, name);
    635 
    636     if(name == 0) return -1;
    637     if(strlen(name) > 256) return -1;
    638 
    639     if ((name[0] == '/') && ((fd = _open_lib(name)) >= 0))
    640         return fd;
    641 
    642     for (path = ldpaths; *path; path++) {
    643         n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
    644         if (n < 0 || n >= (int)sizeof(buf)) {
    645             WARN("Ignoring very long library path: %s/%s\n", *path, name);
    646             continue;
    647         }
    648         if ((fd = _open_lib(buf)) >= 0)
    649             return fd;
    650     }
    651     for (path = sopaths; *path; path++) {
    652         n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
    653         if (n < 0 || n >= (int)sizeof(buf)) {
    654             WARN("Ignoring very long library path: %s/%s\n", *path, name);
    655             continue;
    656         }
    657         if ((fd = _open_lib(buf)) >= 0)
    658             return fd;
    659     }
    660 
    661     return -1;
    662 }
    663 
    664 /* temporary space for holding the first page of the shared lib
    665  * which contains the elf header (with the pht). */
    666 static unsigned char __header[PAGE_SIZE];
    667 
    668 typedef struct {
    669     long mmap_addr;
    670     char tag[4]; /* 'P', 'R', 'E', ' ' */
    671 } prelink_info_t;
    672 
    673 /* Returns the requested base address if the library is prelinked,
    674  * and 0 otherwise.  */
    675 static unsigned long
    676 is_prelinked(int fd, const char *name)
    677 {
    678     off_t sz;
    679     prelink_info_t info;
    680 
    681     sz = lseek(fd, -sizeof(prelink_info_t), SEEK_END);
    682     if (sz < 0) {
    683         DL_ERR("lseek() failed!");
    684         return 0;
    685     }
    686 
    687     if (read(fd, &info, sizeof(info)) != sizeof(info)) {
    688         WARN("Could not read prelink_info_t structure for `%s`\n", name);
    689         return 0;
    690     }
    691 
    692     if (strncmp(info.tag, "PRE ", 4)) {
    693         WARN("`%s` is not a prelinked library\n", name);
    694         return 0;
    695     }
    696 
    697     return (unsigned long)info.mmap_addr;
    698 }
    699 
    700 /* verify_elf_object
    701  *      Verifies if the object @ base is a valid ELF object
    702  *
    703  * Args:
    704  *
    705  * Returns:
    706  *       0 on success
    707  *      -1 if no valid ELF object is found @ base.
    708  */
    709 static int
    710 verify_elf_object(void *base, const char *name)
    711 {
    712     Elf32_Ehdr *hdr = (Elf32_Ehdr *) base;
    713 
    714     if (hdr->e_ident[EI_MAG0] != ELFMAG0) return -1;
    715     if (hdr->e_ident[EI_MAG1] != ELFMAG1) return -1;
    716     if (hdr->e_ident[EI_MAG2] != ELFMAG2) return -1;
    717     if (hdr->e_ident[EI_MAG3] != ELFMAG3) return -1;
    718 
    719     /* TODO: Should we verify anything else in the header? */
    720 
    721     return 0;
    722 }
    723 
    724 
    725 /* get_lib_extents
    726  *      Retrieves the base (*base) address where the ELF object should be
    727  *      mapped and its overall memory size (*total_sz).
    728  *
    729  * Args:
    730  *      fd: Opened file descriptor for the library
    731  *      name: The name of the library
    732  *      _hdr: Pointer to the header page of the library
    733  *      total_sz: Total size of the memory that should be allocated for
    734  *                this library
    735  *
    736  * Returns:
    737  *      -1 if there was an error while trying to get the lib extents.
    738  *         The possible reasons are:
    739  *             - Could not determine if the library was prelinked.
    740  *             - The library provided is not a valid ELF object
    741  *       0 if the library did not request a specific base offset (normal
    742  *         for non-prelinked libs)
    743  *     > 0 if the library requests a specific address to be mapped to.
    744  *         This indicates a pre-linked library.
    745  */
    746 static unsigned
    747 get_lib_extents(int fd, const char *name, void *__hdr, unsigned *total_sz)
    748 {
    749     unsigned req_base;
    750     unsigned min_vaddr = 0xffffffff;
    751     unsigned max_vaddr = 0;
    752     unsigned char *_hdr = (unsigned char *)__hdr;
    753     Elf32_Ehdr *ehdr = (Elf32_Ehdr *)_hdr;
    754     Elf32_Phdr *phdr;
    755     int cnt;
    756 
    757     TRACE("[ %5d Computing extents for '%s'. ]\n", pid, name);
    758     if (verify_elf_object(_hdr, name) < 0) {
    759         DL_ERR("%5d - %s is not a valid ELF object", pid, name);
    760         return (unsigned)-1;
    761     }
    762 
    763     req_base = (unsigned) is_prelinked(fd, name);
    764     if (req_base == (unsigned)-1)
    765         return -1;
    766     else if (req_base != 0) {
    767         TRACE("[ %5d - Prelinked library '%s' requesting base @ 0x%08x ]\n",
    768               pid, name, req_base);
    769     } else {
    770         TRACE("[ %5d - Non-prelinked library '%s' found. ]\n", pid, name);
    771     }
    772 
    773     phdr = (Elf32_Phdr *)(_hdr + ehdr->e_phoff);
    774 
    775     /* find the min/max p_vaddrs from all the PT_LOAD segments so we can
    776      * get the range. */
    777     for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
    778         if (phdr->p_type == PT_LOAD) {
    779             if ((phdr->p_vaddr + phdr->p_memsz) > max_vaddr)
    780                 max_vaddr = phdr->p_vaddr + phdr->p_memsz;
    781             if (phdr->p_vaddr < min_vaddr)
    782                 min_vaddr = phdr->p_vaddr;
    783         }
    784     }
    785 
    786     if ((min_vaddr == 0xffffffff) && (max_vaddr == 0)) {
    787         DL_ERR("%5d - No loadable segments found in %s.", pid, name);
    788         return (unsigned)-1;
    789     }
    790 
    791     /* truncate min_vaddr down to page boundary */
    792     min_vaddr &= ~PAGE_MASK;
    793 
    794     /* round max_vaddr up to the next page */
    795     max_vaddr = (max_vaddr + PAGE_SIZE - 1) & ~PAGE_MASK;
    796 
    797     *total_sz = (max_vaddr - min_vaddr);
    798     return (unsigned)req_base;
    799 }
    800 
    801 /* alloc_mem_region
    802  *
    803  *     This function reserves a chunk of memory to be used for mapping in
    804  *     the shared library. We reserve the entire memory region here, and
    805  *     then the rest of the linker will relocate the individual loadable
    806  *     segments into the correct locations within this memory range.
    807  *
    808  * Args:
    809  *     si->base: The requested base of the allocation. If 0, a sane one will be
    810  *               chosen in the range LIBBASE <= base < LIBLAST.
    811  *     si->size: The size of the allocation.
    812  *
    813  * Returns:
    814  *     -1 on failure, and 0 on success.  On success, si->base will contain
    815  *     the virtual address at which the library will be mapped.
    816  */
    817 
    818 static int reserve_mem_region(soinfo *si)
    819 {
    820     void *base = mmap((void *)si->base, si->size, PROT_READ | PROT_EXEC,
    821                       MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
    822     if (base == MAP_FAILED) {
    823         DL_ERR("%5d can NOT map (%sprelinked) library '%s' at 0x%08x "
    824               "as requested, will try general pool: %d (%s)",
    825               pid, (si->base ? "" : "non-"), si->name, si->base,
    826               errno, strerror(errno));
    827         return -1;
    828     } else if (base != (void *)si->base) {
    829         DL_ERR("OOPS: %5d %sprelinked library '%s' mapped at 0x%08x, "
    830               "not at 0x%08x", pid, (si->base ? "" : "non-"),
    831               si->name, (unsigned)base, si->base);
    832         munmap(base, si->size);
    833         return -1;
    834     }
    835     return 0;
    836 }
    837 
    838 static int
    839 alloc_mem_region(soinfo *si)
    840 {
    841     if (si->base) {
    842         /* Attempt to mmap a prelinked library. */
    843         si->ba_index = -1;
    844         return reserve_mem_region(si);
    845     }
    846 
    847     /* This is not a prelinked library, so we attempt to allocate space
    848        for it from the buddy allocator, which manages the area between
    849        LIBBASE and LIBLAST.
    850     */
    851     si->ba_index = ba_allocate(&ba_nonprelink, si->size);
    852     if(si->ba_index >= 0) {
    853         si->base = ba_start_addr(&ba_nonprelink, si->ba_index);
    854         PRINT("%5d mapping library '%s' at %08x (index %d) " \
    855               "through buddy allocator.\n",
    856               pid, si->name, si->base, si->ba_index);
    857         if (reserve_mem_region(si) < 0) {
    858             ba_free(&ba_nonprelink, si->ba_index);
    859             si->ba_index = -1;
    860             si->base = 0;
    861             goto err;
    862         }
    863         return 0;
    864     }
    865 
    866 err:
    867     DL_ERR("OOPS: %5d cannot map library '%s'. no vspace available.",
    868           pid, si->name);
    869     return -1;
    870 }
    871 
    872 #define MAYBE_MAP_FLAG(x,from,to)    (((x) & (from)) ? (to) : 0)
    873 #define PFLAGS_TO_PROT(x)            (MAYBE_MAP_FLAG((x), PF_X, PROT_EXEC) | \
    874                                       MAYBE_MAP_FLAG((x), PF_R, PROT_READ) | \
    875                                       MAYBE_MAP_FLAG((x), PF_W, PROT_WRITE))
    876 /* load_segments
    877  *
    878  *     This function loads all the loadable (PT_LOAD) segments into memory
    879  *     at their appropriate memory offsets off the base address.
    880  *
    881  * Args:
    882  *     fd: Open file descriptor to the library to load.
    883  *     header: Pointer to a header page that contains the ELF header.
    884  *             This is needed since we haven't mapped in the real file yet.
    885  *     si: ptr to soinfo struct describing the shared object.
    886  *
    887  * Returns:
    888  *     0 on success, -1 on failure.
    889  */
    890 static int
    891 load_segments(int fd, void *header, soinfo *si)
    892 {
    893     Elf32_Ehdr *ehdr = (Elf32_Ehdr *)header;
    894     Elf32_Phdr *phdr = (Elf32_Phdr *)((unsigned char *)header + ehdr->e_phoff);
    895     unsigned char *base = (unsigned char *)si->base;
    896     int cnt;
    897     unsigned len;
    898     unsigned char *tmp;
    899     unsigned char *pbase;
    900     unsigned char *extra_base;
    901     unsigned extra_len;
    902     unsigned total_sz = 0;
    903 
    904     si->wrprotect_start = 0xffffffff;
    905     si->wrprotect_end = 0;
    906 
    907     TRACE("[ %5d - Begin loading segments for '%s' @ 0x%08x ]\n",
    908           pid, si->name, (unsigned)si->base);
    909     /* Now go through all the PT_LOAD segments and map them into memory
    910      * at the appropriate locations. */
    911     for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
    912         if (phdr->p_type == PT_LOAD) {
    913             DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
    914             /* we want to map in the segment on a page boundary */
    915             tmp = base + (phdr->p_vaddr & (~PAGE_MASK));
    916             /* add the # of bytes we masked off above to the total length. */
    917             len = phdr->p_filesz + (phdr->p_vaddr & PAGE_MASK);
    918 
    919             TRACE("[ %d - Trying to load segment from '%s' @ 0x%08x "
    920                   "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x ]\n", pid, si->name,
    921                   (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
    922             pbase = mmap(tmp, len, PFLAGS_TO_PROT(phdr->p_flags),
    923                          MAP_PRIVATE | MAP_FIXED, fd,
    924                          phdr->p_offset & (~PAGE_MASK));
    925             if (pbase == MAP_FAILED) {
    926                 DL_ERR("%d failed to map segment from '%s' @ 0x%08x (0x%08x). "
    927                       "p_vaddr=0x%08x p_offset=0x%08x", pid, si->name,
    928                       (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
    929                 goto fail;
    930             }
    931 
    932             /* If 'len' didn't end on page boundary, and it's a writable
    933              * segment, zero-fill the rest. */
    934             if ((len & PAGE_MASK) && (phdr->p_flags & PF_W))
    935                 memset((void *)(pbase + len), 0, PAGE_SIZE - (len & PAGE_MASK));
    936 
    937             /* Check to see if we need to extend the map for this segment to
    938              * cover the diff between filesz and memsz (i.e. for bss).
    939              *
    940              *  base           _+---------------------+  page boundary
    941              *                  .                     .
    942              *                  |                     |
    943              *                  .                     .
    944              *  pbase          _+---------------------+  page boundary
    945              *                  |                     |
    946              *                  .                     .
    947              *  base + p_vaddr _|                     |
    948              *                  . \          \        .
    949              *                  . | filesz   |        .
    950              *  pbase + len    _| /          |        |
    951              *     <0 pad>      .            .        .
    952              *  extra_base     _+------------|--------+  page boundary
    953              *               /  .            .        .
    954              *               |  .            .        .
    955              *               |  +------------|--------+  page boundary
    956              *  extra_len->  |  |            |        |
    957              *               |  .            | memsz  .
    958              *               |  .            |        .
    959              *               \ _|            /        |
    960              *                  .                     .
    961              *                  |                     |
    962              *                 _+---------------------+  page boundary
    963              */
    964             tmp = (unsigned char *)(((unsigned)pbase + len + PAGE_SIZE - 1) &
    965                                     (~PAGE_MASK));
    966             if (tmp < (base + phdr->p_vaddr + phdr->p_memsz)) {
    967                 extra_len = base + phdr->p_vaddr + phdr->p_memsz - tmp;
    968                 TRACE("[ %5d - Need to extend segment from '%s' @ 0x%08x "
    969                       "(0x%08x) ]\n", pid, si->name, (unsigned)tmp, extra_len);
    970                 /* map in the extra page(s) as anonymous into the range.
    971                  * This is probably not necessary as we already mapped in
    972                  * the entire region previously, but we just want to be
    973                  * sure. This will also set the right flags on the region
    974                  * (though we can probably accomplish the same thing with
    975                  * mprotect).
    976                  */
    977                 extra_base = mmap((void *)tmp, extra_len,
    978                                   PFLAGS_TO_PROT(phdr->p_flags),
    979                                   MAP_PRIVATE | MAP_FIXED | MAP_ANONYMOUS,
    980                                   -1, 0);
    981                 if (extra_base == MAP_FAILED) {
    982                     DL_ERR("[ %5d - failed to extend segment from '%s' @ 0x%08x"
    983                            " (0x%08x) ]", pid, si->name, (unsigned)tmp,
    984                           extra_len);
    985                     goto fail;
    986                 }
    987                 /* TODO: Check if we need to memset-0 this region.
    988                  * Anonymous mappings are zero-filled copy-on-writes, so we
    989                  * shouldn't need to. */
    990                 TRACE("[ %5d - Segment from '%s' extended @ 0x%08x "
    991                       "(0x%08x)\n", pid, si->name, (unsigned)extra_base,
    992                       extra_len);
    993             }
    994             /* set the len here to show the full extent of the segment we
    995              * just loaded, mostly for debugging */
    996             len = (((unsigned)base + phdr->p_vaddr + phdr->p_memsz +
    997                     PAGE_SIZE - 1) & (~PAGE_MASK)) - (unsigned)pbase;
    998             TRACE("[ %5d - Successfully loaded segment from '%s' @ 0x%08x "
    999                   "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x\n", pid, si->name,
   1000                   (unsigned)pbase, len, phdr->p_vaddr, phdr->p_offset);
   1001             total_sz += len;
   1002             /* Make the section writable just in case we'll have to write to
   1003              * it during relocation (i.e. text segment). However, we will
   1004              * remember what range of addresses should be write protected.
   1005              *
   1006              */
   1007             if (!(phdr->p_flags & PF_W)) {
   1008                 if ((unsigned)pbase < si->wrprotect_start)
   1009                     si->wrprotect_start = (unsigned)pbase;
   1010                 if (((unsigned)pbase + len) > si->wrprotect_end)
   1011                     si->wrprotect_end = (unsigned)pbase + len;
   1012                 mprotect(pbase, len,
   1013                          PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
   1014             }
   1015         } else if (phdr->p_type == PT_DYNAMIC) {
   1016             DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
   1017             /* this segment contains the dynamic linking information */
   1018             si->dynamic = (unsigned *)(base + phdr->p_vaddr);
   1019         } else {
   1020 #ifdef ANDROID_ARM_LINKER
   1021             if (phdr->p_type == PT_ARM_EXIDX) {
   1022                 DEBUG_DUMP_PHDR(phdr, "PT_ARM_EXIDX", pid);
   1023                 /* exidx entries (used for stack unwinding) are 8 bytes each.
   1024                  */
   1025                 si->ARM_exidx = (unsigned *)phdr->p_vaddr;
   1026                 si->ARM_exidx_count = phdr->p_memsz / 8;
   1027             }
   1028 #endif
   1029         }
   1030 
   1031     }
   1032 
   1033     /* Sanity check */
   1034     if (total_sz > si->size) {
   1035         DL_ERR("%5d - Total length (0x%08x) of mapped segments from '%s' is "
   1036               "greater than what was allocated (0x%08x). THIS IS BAD!",
   1037               pid, total_sz, si->name, si->size);
   1038         goto fail;
   1039     }
   1040 
   1041     TRACE("[ %5d - Finish loading segments for '%s' @ 0x%08x. "
   1042           "Total memory footprint: 0x%08x bytes ]\n", pid, si->name,
   1043           (unsigned)si->base, si->size);
   1044     return 0;
   1045 
   1046 fail:
   1047     /* We can just blindly unmap the entire region even though some things
   1048      * were mapped in originally with anonymous and others could have been
   1049      * been mapped in from the file before we failed. The kernel will unmap
   1050      * all the pages in the range, irrespective of how they got there.
   1051      */
   1052     munmap((void *)si->base, si->size);
   1053     si->flags |= FLAG_ERROR;
   1054     return -1;
   1055 }
   1056 
   1057 /* TODO: Implement this to take care of the fact that Android ARM
   1058  * ELF objects shove everything into a single loadable segment that has the
   1059  * write bit set. wr_offset is then used to set non-(data|bss) pages to be
   1060  * non-writable.
   1061  */
   1062 #if 0
   1063 static unsigned
   1064 get_wr_offset(int fd, const char *name, Elf32_Ehdr *ehdr)
   1065 {
   1066     Elf32_Shdr *shdr_start;
   1067     Elf32_Shdr *shdr;
   1068     int shdr_sz = ehdr->e_shnum * sizeof(Elf32_Shdr);
   1069     int cnt;
   1070     unsigned wr_offset = 0xffffffff;
   1071 
   1072     shdr_start = mmap(0, shdr_sz, PROT_READ, MAP_PRIVATE, fd,
   1073                       ehdr->e_shoff & (~PAGE_MASK));
   1074     if (shdr_start == MAP_FAILED) {
   1075         WARN("%5d - Could not read section header info from '%s'. Will not "
   1076              "not be able to determine write-protect offset.\n", pid, name);
   1077         return (unsigned)-1;
   1078     }
   1079 
   1080     for(cnt = 0, shdr = shdr_start; cnt < ehdr->e_shnum; ++cnt, ++shdr) {
   1081         if ((shdr->sh_type != SHT_NULL) && (shdr->sh_flags & SHF_WRITE) &&
   1082             (shdr->sh_addr < wr_offset)) {
   1083             wr_offset = shdr->sh_addr;
   1084         }
   1085     }
   1086 
   1087     munmap(shdr_start, shdr_sz);
   1088     return wr_offset;
   1089 }
   1090 #endif
   1091 
   1092 static soinfo *
   1093 load_library(const char *name)
   1094 {
   1095     int fd = open_library(name);
   1096     int cnt;
   1097     unsigned ext_sz;
   1098     unsigned req_base;
   1099     const char *bname;
   1100     soinfo *si = NULL;
   1101     Elf32_Ehdr *hdr;
   1102 
   1103     if(fd == -1) {
   1104         DL_ERR("Library '%s' not found", name);
   1105         return NULL;
   1106     }
   1107 
   1108     /* We have to read the ELF header to figure out what to do with this image
   1109      */
   1110     if (lseek(fd, 0, SEEK_SET) < 0) {
   1111         DL_ERR("lseek() failed!");
   1112         goto fail;
   1113     }
   1114 
   1115     if ((cnt = read(fd, &__header[0], PAGE_SIZE)) < 0) {
   1116         DL_ERR("read() failed!");
   1117         goto fail;
   1118     }
   1119 
   1120     /* Parse the ELF header and get the size of the memory footprint for
   1121      * the library */
   1122     req_base = get_lib_extents(fd, name, &__header[0], &ext_sz);
   1123     if (req_base == (unsigned)-1)
   1124         goto fail;
   1125     TRACE("[ %5d - '%s' (%s) wants base=0x%08x sz=0x%08x ]\n", pid, name,
   1126           (req_base ? "prelinked" : "not pre-linked"), req_base, ext_sz);
   1127 
   1128     /* Now configure the soinfo struct where we'll store all of our data
   1129      * for the ELF object. If the loading fails, we waste the entry, but
   1130      * same thing would happen if we failed during linking. Configuring the
   1131      * soinfo struct here is a lot more convenient.
   1132      */
   1133     bname = strrchr(name, '/');
   1134     si = alloc_info(bname ? bname + 1 : name);
   1135     if (si == NULL)
   1136         goto fail;
   1137 
   1138     /* Carve out a chunk of memory where we will map in the individual
   1139      * segments */
   1140     si->base = req_base;
   1141     si->size = ext_sz;
   1142     si->flags = 0;
   1143     si->entry = 0;
   1144     si->dynamic = (unsigned *)-1;
   1145     if (alloc_mem_region(si) < 0)
   1146         goto fail;
   1147 
   1148     TRACE("[ %5d allocated memory for %s @ %p (0x%08x) ]\n",
   1149           pid, name, (void *)si->base, (unsigned) ext_sz);
   1150 
   1151     /* Now actually load the library's segments into right places in memory */
   1152     if (load_segments(fd, &__header[0], si) < 0) {
   1153         if (si->ba_index >= 0) {
   1154             ba_free(&ba_nonprelink, si->ba_index);
   1155             si->ba_index = -1;
   1156         }
   1157         goto fail;
   1158     }
   1159 
   1160     /* this might not be right. Technically, we don't even need this info
   1161      * once we go through 'load_segments'. */
   1162     hdr = (Elf32_Ehdr *)si->base;
   1163     si->phdr = (Elf32_Phdr *)((unsigned char *)si->base + hdr->e_phoff);
   1164     si->phnum = hdr->e_phnum;
   1165     /**/
   1166 
   1167     close(fd);
   1168     return si;
   1169 
   1170 fail:
   1171     if (si) free_info(si);
   1172     close(fd);
   1173     return NULL;
   1174 }
   1175 
   1176 static soinfo *
   1177 init_library(soinfo *si)
   1178 {
   1179     unsigned wr_offset = 0xffffffff;
   1180 
   1181     /* At this point we know that whatever is loaded @ base is a valid ELF
   1182      * shared library whose segments are properly mapped in. */
   1183     TRACE("[ %5d init_library base=0x%08x sz=0x%08x name='%s') ]\n",
   1184           pid, si->base, si->size, si->name);
   1185 
   1186     if (si->base < LIBBASE || si->base >= LIBLAST)
   1187         si->flags |= FLAG_PRELINKED;
   1188 
   1189     if(link_image(si, wr_offset)) {
   1190             /* We failed to link.  However, we can only restore libbase
   1191             ** if no additional libraries have moved it since we updated it.
   1192             */
   1193         munmap((void *)si->base, si->size);
   1194         return NULL;
   1195     }
   1196 
   1197     return si;
   1198 }
   1199 
   1200 soinfo *find_library(const char *name)
   1201 {
   1202     soinfo *si;
   1203     const char *bname;
   1204 
   1205 #if ALLOW_SYMBOLS_FROM_MAIN
   1206     if (name == NULL)
   1207         return somain;
   1208 #else
   1209     if (name == NULL)
   1210         return NULL;
   1211 #endif
   1212 
   1213     bname = strrchr(name, '/');
   1214     bname = bname ? bname + 1 : name;
   1215 
   1216     for(si = solist; si != 0; si = si->next){
   1217         if(!strcmp(bname, si->name)) {
   1218             if(si->flags & FLAG_ERROR) {
   1219                 DL_ERR("%5d '%s' failed to load previously", pid, bname);
   1220                 return NULL;
   1221             }
   1222             if(si->flags & FLAG_LINKED) return si;
   1223             DL_ERR("OOPS: %5d recursive link to '%s'", pid, si->name);
   1224             return NULL;
   1225         }
   1226     }
   1227 
   1228     TRACE("[ %5d '%s' has not been loaded yet.  Locating...]\n", pid, name);
   1229     si = load_library(name);
   1230     if(si == NULL)
   1231         return NULL;
   1232     return init_library(si);
   1233 }
   1234 
   1235 /* TODO:
   1236  *   notify gdb of unload
   1237  *   for non-prelinked libraries, find a way to decrement libbase
   1238  */
   1239 static void call_destructors(soinfo *si);
   1240 unsigned unload_library(soinfo *si)
   1241 {
   1242     unsigned *d;
   1243     if (si->refcount == 1) {
   1244         TRACE("%5d unloading '%s'\n", pid, si->name);
   1245         call_destructors(si);
   1246 
   1247         for(d = si->dynamic; *d; d += 2) {
   1248             if(d[0] == DT_NEEDED){
   1249                 soinfo *lsi = (soinfo *)d[1];
   1250                 d[1] = 0;
   1251                 if (validate_soinfo(lsi)) {
   1252                     TRACE("%5d %s needs to unload %s\n", pid,
   1253                           si->name, lsi->name);
   1254                     unload_library(lsi);
   1255                 }
   1256                 else
   1257                     DL_ERR("%5d %s: could not unload dependent library",
   1258                            pid, si->name);
   1259             }
   1260         }
   1261 
   1262         munmap((char *)si->base, si->size);
   1263         if (si->ba_index >= 0) {
   1264             PRINT("%5d releasing library '%s' address space at %08x "\
   1265                   "through buddy allocator.\n",
   1266                   pid, si->name, si->base);
   1267             ba_free(&ba_nonprelink, si->ba_index);
   1268         }
   1269         notify_gdb_of_unload(si);
   1270         free_info(si);
   1271         si->refcount = 0;
   1272     }
   1273     else {
   1274         si->refcount--;
   1275         PRINT("%5d not unloading '%s', decrementing refcount to %d\n",
   1276               pid, si->name, si->refcount);
   1277     }
   1278     return si->refcount;
   1279 }
   1280 
   1281 /* TODO: don't use unsigned for addrs below. It works, but is not
   1282  * ideal. They should probably be either uint32_t, Elf32_Addr, or unsigned
   1283  * long.
   1284  */
   1285 static int reloc_library(soinfo *si, Elf32_Rel *rel, unsigned count)
   1286 {
   1287     Elf32_Sym *symtab = si->symtab;
   1288     const char *strtab = si->strtab;
   1289     Elf32_Sym *s;
   1290     unsigned base;
   1291     Elf32_Rel *start = rel;
   1292     unsigned idx;
   1293 
   1294     for (idx = 0; idx < count; ++idx) {
   1295         unsigned type = ELF32_R_TYPE(rel->r_info);
   1296         unsigned sym = ELF32_R_SYM(rel->r_info);
   1297         unsigned reloc = (unsigned)(rel->r_offset + si->base);
   1298         unsigned sym_addr = 0;
   1299         char *sym_name = NULL;
   1300 
   1301         DEBUG("%5d Processing '%s' relocation at index %d\n", pid,
   1302               si->name, idx);
   1303         if(sym != 0) {
   1304             sym_name = (char *)(strtab + symtab[sym].st_name);
   1305             s = _do_lookup(si, sym_name, &base);
   1306             if(s == NULL) {
   1307                 /* We only allow an undefined symbol if this is a weak
   1308                    reference..   */
   1309                 s = &symtab[sym];
   1310                 if (ELF32_ST_BIND(s->st_info) != STB_WEAK) {
   1311                     DL_ERR("%5d cannot locate '%s'...\n", pid, sym_name);
   1312                     return -1;
   1313                 }
   1314 
   1315                 /* IHI0044C AAELF 4.5.1.1:
   1316 
   1317                    Libraries are not searched to resolve weak references.
   1318                    It is not an error for a weak reference to remain
   1319                    unsatisfied.
   1320 
   1321                    During linking, the value of an undefined weak reference is:
   1322                    - Zero if the relocation type is absolute
   1323                    - The address of the place if the relocation is pc-relative
   1324                    - The address of nominial base address if the relocation
   1325                      type is base-relative.
   1326                   */
   1327 
   1328                 switch (type) {
   1329 #if defined(ANDROID_ARM_LINKER)
   1330                 case R_ARM_JUMP_SLOT:
   1331                 case R_ARM_GLOB_DAT:
   1332                 case R_ARM_ABS32:
   1333                 case R_ARM_RELATIVE:    /* Don't care. */
   1334                 case R_ARM_NONE:        /* Don't care. */
   1335 #elif defined(ANDROID_X86_LINKER)
   1336                 case R_386_JUMP_SLOT:
   1337                 case R_386_GLOB_DAT:
   1338                 case R_386_32:
   1339                 case R_386_RELATIVE:    /* Dont' care. */
   1340 #endif /* ANDROID_*_LINKER */
   1341                     /* sym_addr was initialized to be zero above or relocation
   1342                        code below does not care about value of sym_addr.
   1343                        No need to do anything.  */
   1344                     break;
   1345 
   1346 #if defined(ANDROID_X86_LINKER)
   1347                 case R_386_PC32:
   1348                     sym_addr = reloc;
   1349                     break;
   1350 #endif /* ANDROID_X86_LINKER */
   1351 
   1352 #if defined(ANDROID_ARM_LINKER)
   1353                 case R_ARM_COPY:
   1354                     /* Fall through.  Can't really copy if weak symbol is
   1355                        not found in run-time.  */
   1356 #endif /* ANDROID_ARM_LINKER */
   1357                 default:
   1358                     DL_ERR("%5d unknown weak reloc type %d @ %p (%d)\n",
   1359                                  pid, type, rel, (int) (rel - start));
   1360                     return -1;
   1361                 }
   1362             } else {
   1363                 /* We got a definition.  */
   1364 #if 0
   1365             if((base == 0) && (si->base != 0)){
   1366                     /* linking from libraries to main image is bad */
   1367                 DL_ERR("%5d cannot locate '%s'...",
   1368                        pid, strtab + symtab[sym].st_name);
   1369                 return -1;
   1370             }
   1371 #endif
   1372                 sym_addr = (unsigned)(s->st_value + base);
   1373 	    }
   1374             COUNT_RELOC(RELOC_SYMBOL);
   1375         } else {
   1376             s = NULL;
   1377         }
   1378 
   1379 /* TODO: This is ugly. Split up the relocations by arch into
   1380  * different files.
   1381  */
   1382         switch(type){
   1383 #if defined(ANDROID_ARM_LINKER)
   1384         case R_ARM_JUMP_SLOT:
   1385             COUNT_RELOC(RELOC_ABSOLUTE);
   1386             MARK(rel->r_offset);
   1387             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1388                        reloc, sym_addr, sym_name);
   1389             *((unsigned*)reloc) = sym_addr;
   1390             break;
   1391         case R_ARM_GLOB_DAT:
   1392             COUNT_RELOC(RELOC_ABSOLUTE);
   1393             MARK(rel->r_offset);
   1394             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1395                        reloc, sym_addr, sym_name);
   1396             *((unsigned*)reloc) = sym_addr;
   1397             break;
   1398         case R_ARM_ABS32:
   1399             COUNT_RELOC(RELOC_ABSOLUTE);
   1400             MARK(rel->r_offset);
   1401             TRACE_TYPE(RELO, "%5d RELO ABS %08x <- %08x %s\n", pid,
   1402                        reloc, sym_addr, sym_name);
   1403             *((unsigned*)reloc) += sym_addr;
   1404             break;
   1405         case R_ARM_REL32:
   1406             COUNT_RELOC(RELOC_RELATIVE);
   1407             MARK(rel->r_offset);
   1408             TRACE_TYPE(RELO, "%5d RELO REL32 %08x <- %08x - %08x %s\n", pid,
   1409                        reloc, sym_addr, rel->r_offset, sym_name);
   1410             *((unsigned*)reloc) += sym_addr - rel->r_offset;
   1411             break;
   1412 #elif defined(ANDROID_X86_LINKER)
   1413         case R_386_JUMP_SLOT:
   1414             COUNT_RELOC(RELOC_ABSOLUTE);
   1415             MARK(rel->r_offset);
   1416             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1417                        reloc, sym_addr, sym_name);
   1418             *((unsigned*)reloc) = sym_addr;
   1419             break;
   1420         case R_386_GLOB_DAT:
   1421             COUNT_RELOC(RELOC_ABSOLUTE);
   1422             MARK(rel->r_offset);
   1423             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1424                        reloc, sym_addr, sym_name);
   1425             *((unsigned*)reloc) = sym_addr;
   1426             break;
   1427 #endif /* ANDROID_*_LINKER */
   1428 
   1429 #if defined(ANDROID_ARM_LINKER)
   1430         case R_ARM_RELATIVE:
   1431 #elif defined(ANDROID_X86_LINKER)
   1432         case R_386_RELATIVE:
   1433 #endif /* ANDROID_*_LINKER */
   1434             COUNT_RELOC(RELOC_RELATIVE);
   1435             MARK(rel->r_offset);
   1436             if(sym){
   1437                 DL_ERR("%5d odd RELATIVE form...", pid);
   1438                 return -1;
   1439             }
   1440             TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", pid,
   1441                        reloc, si->base);
   1442             *((unsigned*)reloc) += si->base;
   1443             break;
   1444 
   1445 #if defined(ANDROID_X86_LINKER)
   1446         case R_386_32:
   1447             COUNT_RELOC(RELOC_RELATIVE);
   1448             MARK(rel->r_offset);
   1449 
   1450             TRACE_TYPE(RELO, "%5d RELO R_386_32 %08x <- +%08x %s\n", pid,
   1451                        reloc, sym_addr, sym_name);
   1452             *((unsigned *)reloc) += (unsigned)sym_addr;
   1453             break;
   1454 
   1455         case R_386_PC32:
   1456             COUNT_RELOC(RELOC_RELATIVE);
   1457             MARK(rel->r_offset);
   1458             TRACE_TYPE(RELO, "%5d RELO R_386_PC32 %08x <- "
   1459                        "+%08x (%08x - %08x) %s\n", pid, reloc,
   1460                        (sym_addr - reloc), sym_addr, reloc, sym_name);
   1461             *((unsigned *)reloc) += (unsigned)(sym_addr - reloc);
   1462             break;
   1463 #endif /* ANDROID_X86_LINKER */
   1464 
   1465 #ifdef ANDROID_ARM_LINKER
   1466         case R_ARM_COPY:
   1467             COUNT_RELOC(RELOC_COPY);
   1468             MARK(rel->r_offset);
   1469             TRACE_TYPE(RELO, "%5d RELO %08x <- %d @ %08x %s\n", pid,
   1470                        reloc, s->st_size, sym_addr, sym_name);
   1471             memcpy((void*)reloc, (void*)sym_addr, s->st_size);
   1472             break;
   1473         case R_ARM_NONE:
   1474             break;
   1475 #endif /* ANDROID_ARM_LINKER */
   1476 
   1477         default:
   1478             DL_ERR("%5d unknown reloc type %d @ %p (%d)",
   1479                   pid, type, rel, (int) (rel - start));
   1480             return -1;
   1481         }
   1482         rel++;
   1483     }
   1484     return 0;
   1485 }
   1486 
   1487 #if defined(ANDROID_SH_LINKER)
   1488 static int reloc_library_a(soinfo *si, Elf32_Rela *rela, unsigned count)
   1489 {
   1490     Elf32_Sym *symtab = si->symtab;
   1491     const char *strtab = si->strtab;
   1492     Elf32_Sym *s;
   1493     unsigned base;
   1494     Elf32_Rela *start = rela;
   1495     unsigned idx;
   1496 
   1497     for (idx = 0; idx < count; ++idx) {
   1498         unsigned type = ELF32_R_TYPE(rela->r_info);
   1499         unsigned sym = ELF32_R_SYM(rela->r_info);
   1500         unsigned reloc = (unsigned)(rela->r_offset + si->base);
   1501         unsigned sym_addr = 0;
   1502         char *sym_name = NULL;
   1503 
   1504         DEBUG("%5d Processing '%s' relocation at index %d\n", pid,
   1505               si->name, idx);
   1506         if(sym != 0) {
   1507             sym_name = (char *)(strtab + symtab[sym].st_name);
   1508             s = _do_lookup(si, sym_name, &base);
   1509             if(s == 0) {
   1510                 DL_ERR("%5d cannot locate '%s'...", pid, sym_name);
   1511                 return -1;
   1512             }
   1513 #if 0
   1514             if((base == 0) && (si->base != 0)){
   1515                     /* linking from libraries to main image is bad */
   1516                 DL_ERR("%5d cannot locate '%s'...",
   1517                        pid, strtab + symtab[sym].st_name);
   1518                 return -1;
   1519             }
   1520 #endif
   1521             if ((s->st_shndx == SHN_UNDEF) && (s->st_value != 0)) {
   1522                 DL_ERR("%5d In '%s', shndx=%d && value=0x%08x. We do not "
   1523                       "handle this yet", pid, si->name, s->st_shndx,
   1524                       s->st_value);
   1525                 return -1;
   1526             }
   1527             sym_addr = (unsigned)(s->st_value + base);
   1528             COUNT_RELOC(RELOC_SYMBOL);
   1529         } else {
   1530             s = 0;
   1531         }
   1532 
   1533 /* TODO: This is ugly. Split up the relocations by arch into
   1534  * different files.
   1535  */
   1536         switch(type){
   1537         case R_SH_JUMP_SLOT:
   1538             COUNT_RELOC(RELOC_ABSOLUTE);
   1539             MARK(rela->r_offset);
   1540             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1541                        reloc, sym_addr, sym_name);
   1542             *((unsigned*)reloc) = sym_addr;
   1543             break;
   1544         case R_SH_GLOB_DAT:
   1545             COUNT_RELOC(RELOC_ABSOLUTE);
   1546             MARK(rela->r_offset);
   1547             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1548                        reloc, sym_addr, sym_name);
   1549             *((unsigned*)reloc) = sym_addr;
   1550             break;
   1551         case R_SH_DIR32:
   1552             COUNT_RELOC(RELOC_ABSOLUTE);
   1553             MARK(rela->r_offset);
   1554             TRACE_TYPE(RELO, "%5d RELO DIR32 %08x <- %08x %s\n", pid,
   1555                        reloc, sym_addr, sym_name);
   1556             *((unsigned*)reloc) += sym_addr;
   1557             break;
   1558         case R_SH_RELATIVE:
   1559             COUNT_RELOC(RELOC_RELATIVE);
   1560             MARK(rela->r_offset);
   1561             if(sym){
   1562                 DL_ERR("%5d odd RELATIVE form...", pid);
   1563                 return -1;
   1564             }
   1565             TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", pid,
   1566                        reloc, si->base);
   1567             *((unsigned*)reloc) += si->base;
   1568             break;
   1569 
   1570         default:
   1571             DL_ERR("%5d unknown reloc type %d @ %p (%d)",
   1572                   pid, type, rela, (int) (rela - start));
   1573             return -1;
   1574         }
   1575         rela++;
   1576     }
   1577     return 0;
   1578 }
   1579 #endif /* ANDROID_SH_LINKER */
   1580 
   1581 
   1582 /* Please read the "Initialization and Termination functions" functions.
   1583  * of the linker design note in bionic/linker/README.TXT to understand
   1584  * what the following code is doing.
   1585  *
   1586  * The important things to remember are:
   1587  *
   1588  *   DT_PREINIT_ARRAY must be called first for executables, and should
   1589  *   not appear in shared libraries.
   1590  *
   1591  *   DT_INIT should be called before DT_INIT_ARRAY if both are present
   1592  *
   1593  *   DT_FINI should be called after DT_FINI_ARRAY if both are present
   1594  *
   1595  *   DT_FINI_ARRAY must be parsed in reverse order.
   1596  */
   1597 
   1598 static void call_array(unsigned *ctor, int count, int reverse)
   1599 {
   1600     int n, inc = 1;
   1601 
   1602     if (reverse) {
   1603         ctor += (count-1);
   1604         inc   = -1;
   1605     }
   1606 
   1607     for(n = count; n > 0; n--) {
   1608         TRACE("[ %5d Looking at %s *0x%08x == 0x%08x ]\n", pid,
   1609               reverse ? "dtor" : "ctor",
   1610               (unsigned)ctor, (unsigned)*ctor);
   1611         void (*func)() = (void (*)()) *ctor;
   1612         ctor += inc;
   1613         if(((int) func == 0) || ((int) func == -1)) continue;
   1614         TRACE("[ %5d Calling func @ 0x%08x ]\n", pid, (unsigned)func);
   1615         func();
   1616     }
   1617 }
   1618 
   1619 static void call_constructors(soinfo *si)
   1620 {
   1621     if (si->flags & FLAG_EXE) {
   1622         TRACE("[ %5d Calling preinit_array @ 0x%08x [%d] for '%s' ]\n",
   1623               pid, (unsigned)si->preinit_array, si->preinit_array_count,
   1624               si->name);
   1625         call_array(si->preinit_array, si->preinit_array_count, 0);
   1626         TRACE("[ %5d Done calling preinit_array for '%s' ]\n", pid, si->name);
   1627     } else {
   1628         if (si->preinit_array) {
   1629             DL_ERR("%5d Shared library '%s' has a preinit_array table @ 0x%08x."
   1630                    " This is INVALID.", pid, si->name,
   1631                    (unsigned)si->preinit_array);
   1632         }
   1633     }
   1634 
   1635     if (si->init_func) {
   1636         TRACE("[ %5d Calling init_func @ 0x%08x for '%s' ]\n", pid,
   1637               (unsigned)si->init_func, si->name);
   1638         si->init_func();
   1639         TRACE("[ %5d Done calling init_func for '%s' ]\n", pid, si->name);
   1640     }
   1641 
   1642     if (si->init_array) {
   1643         TRACE("[ %5d Calling init_array @ 0x%08x [%d] for '%s' ]\n", pid,
   1644               (unsigned)si->init_array, si->init_array_count, si->name);
   1645         call_array(si->init_array, si->init_array_count, 0);
   1646         TRACE("[ %5d Done calling init_array for '%s' ]\n", pid, si->name);
   1647     }
   1648 }
   1649 
   1650 
   1651 static void call_destructors(soinfo *si)
   1652 {
   1653     if (si->fini_array) {
   1654         TRACE("[ %5d Calling fini_array @ 0x%08x [%d] for '%s' ]\n", pid,
   1655               (unsigned)si->fini_array, si->fini_array_count, si->name);
   1656         call_array(si->fini_array, si->fini_array_count, 1);
   1657         TRACE("[ %5d Done calling fini_array for '%s' ]\n", pid, si->name);
   1658     }
   1659 
   1660     if (si->fini_func) {
   1661         TRACE("[ %5d Calling fini_func @ 0x%08x for '%s' ]\n", pid,
   1662               (unsigned)si->fini_func, si->name);
   1663         si->fini_func();
   1664         TRACE("[ %5d Done calling fini_func for '%s' ]\n", pid, si->name);
   1665     }
   1666 }
   1667 
   1668 /* Force any of the closed stdin, stdout and stderr to be associated with
   1669    /dev/null. */
   1670 static int nullify_closed_stdio (void)
   1671 {
   1672     int dev_null, i, status;
   1673     int return_value = 0;
   1674 
   1675     dev_null = open("/dev/null", O_RDWR);
   1676     if (dev_null < 0) {
   1677         DL_ERR("Cannot open /dev/null.");
   1678         return -1;
   1679     }
   1680     TRACE("[ %5d Opened /dev/null file-descriptor=%d]\n", pid, dev_null);
   1681 
   1682     /* If any of the stdio file descriptors is valid and not associated
   1683        with /dev/null, dup /dev/null to it.  */
   1684     for (i = 0; i < 3; i++) {
   1685         /* If it is /dev/null already, we are done. */
   1686         if (i == dev_null)
   1687             continue;
   1688 
   1689         TRACE("[ %5d Nullifying stdio file descriptor %d]\n", pid, i);
   1690         /* The man page of fcntl does not say that fcntl(..,F_GETFL)
   1691            can be interrupted but we do this just to be safe. */
   1692         do {
   1693           status = fcntl(i, F_GETFL);
   1694         } while (status < 0 && errno == EINTR);
   1695 
   1696         /* If file is openned, we are good. */
   1697         if (status >= 0)
   1698           continue;
   1699 
   1700         /* The only error we allow is that the file descriptor does not
   1701            exist, in which case we dup /dev/null to it. */
   1702         if (errno != EBADF) {
   1703             DL_ERR("nullify_stdio: unhandled error %s", strerror(errno));
   1704             return_value = -1;
   1705             continue;
   1706         }
   1707 
   1708         /* Try dupping /dev/null to this stdio file descriptor and
   1709            repeat if there is a signal.  Note that any errors in closing
   1710            the stdio descriptor are lost.  */
   1711         do {
   1712             status = dup2(dev_null, i);
   1713         } while (status < 0 && errno == EINTR);
   1714 
   1715         if (status < 0) {
   1716             DL_ERR("nullify_stdio: dup2 error %s", strerror(errno));
   1717             return_value = -1;
   1718             continue;
   1719         }
   1720     }
   1721 
   1722     /* If /dev/null is not one of the stdio file descriptors, close it. */
   1723     if (dev_null > 2) {
   1724         TRACE("[ %5d Closing /dev/null file-descriptor=%d]\n", pid, dev_null);
   1725         do {
   1726             status = close(dev_null);
   1727         } while (status < 0 && errno == EINTR);
   1728 
   1729         if (status < 0) {
   1730             DL_ERR("nullify_stdio: close error %s", strerror(errno));
   1731             return_value = -1;
   1732         }
   1733     }
   1734 
   1735     return return_value;
   1736 }
   1737 
   1738 static int link_image(soinfo *si, unsigned wr_offset)
   1739 {
   1740     unsigned *d;
   1741     Elf32_Phdr *phdr = si->phdr;
   1742     int phnum = si->phnum;
   1743 
   1744     INFO("[ %5d linking %s ]\n", pid, si->name);
   1745     DEBUG("%5d si->base = 0x%08x si->flags = 0x%08x\n", pid,
   1746           si->base, si->flags);
   1747 
   1748     if (si->flags & FLAG_EXE) {
   1749         /* Locate the needed program segments (DYNAMIC/ARM_EXIDX) for
   1750          * linkage info if this is the executable. If this was a
   1751          * dynamic lib, that would have been done at load time.
   1752          *
   1753          * TODO: It's unfortunate that small pieces of this are
   1754          * repeated from the load_library routine. Refactor this just
   1755          * slightly to reuse these bits.
   1756          */
   1757         si->size = 0;
   1758         for(; phnum > 0; --phnum, ++phdr) {
   1759 #ifdef ANDROID_ARM_LINKER
   1760             if(phdr->p_type == PT_ARM_EXIDX) {
   1761                 /* exidx entries (used for stack unwinding) are 8 bytes each.
   1762                  */
   1763                 si->ARM_exidx = (unsigned *)phdr->p_vaddr;
   1764                 si->ARM_exidx_count = phdr->p_memsz / 8;
   1765             }
   1766 #endif
   1767             if (phdr->p_type == PT_LOAD) {
   1768                 /* For the executable, we use the si->size field only in
   1769                    dl_unwind_find_exidx(), so the meaning of si->size
   1770                    is not the size of the executable; it is the last
   1771                    virtual address of the loadable part of the executable;
   1772                    since si->base == 0 for an executable, we use the
   1773                    range [0, si->size) to determine whether a PC value
   1774                    falls within the executable section.  Of course, if
   1775                    a value is below phdr->p_vaddr, it's not in the
   1776                    executable section, but a) we shouldn't be asking for
   1777                    such a value anyway, and b) if we have to provide
   1778                    an EXIDX for such a value, then the executable's
   1779                    EXIDX is probably the better choice.
   1780                 */
   1781                 DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
   1782                 if (phdr->p_vaddr + phdr->p_memsz > si->size)
   1783                     si->size = phdr->p_vaddr + phdr->p_memsz;
   1784                 /* try to remember what range of addresses should be write
   1785                  * protected */
   1786                 if (!(phdr->p_flags & PF_W)) {
   1787                     unsigned _end;
   1788 
   1789                     if (phdr->p_vaddr < si->wrprotect_start)
   1790                         si->wrprotect_start = phdr->p_vaddr;
   1791                     _end = (((phdr->p_vaddr + phdr->p_memsz + PAGE_SIZE - 1) &
   1792                              (~PAGE_MASK)));
   1793                     if (_end > si->wrprotect_end)
   1794                         si->wrprotect_end = _end;
   1795                 }
   1796             } else if (phdr->p_type == PT_DYNAMIC) {
   1797                 if (si->dynamic != (unsigned *)-1) {
   1798                     DL_ERR("%5d multiple PT_DYNAMIC segments found in '%s'. "
   1799                           "Segment at 0x%08x, previously one found at 0x%08x",
   1800                           pid, si->name, si->base + phdr->p_vaddr,
   1801                           (unsigned)si->dynamic);
   1802                     goto fail;
   1803                 }
   1804                 DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
   1805                 si->dynamic = (unsigned *) (si->base + phdr->p_vaddr);
   1806             }
   1807         }
   1808     }
   1809 
   1810     if (si->dynamic == (unsigned *)-1) {
   1811         DL_ERR("%5d missing PT_DYNAMIC?!", pid);
   1812         goto fail;
   1813     }
   1814 
   1815     DEBUG("%5d dynamic = %p\n", pid, si->dynamic);
   1816 
   1817     /* extract useful information from dynamic section */
   1818     for(d = si->dynamic; *d; d++){
   1819         DEBUG("%5d d = %p, d[0] = 0x%08x d[1] = 0x%08x\n", pid, d, d[0], d[1]);
   1820         switch(*d++){
   1821         case DT_HASH:
   1822             si->nbucket = ((unsigned *) (si->base + *d))[0];
   1823             si->nchain = ((unsigned *) (si->base + *d))[1];
   1824             si->bucket = (unsigned *) (si->base + *d + 8);
   1825             si->chain = (unsigned *) (si->base + *d + 8 + si->nbucket * 4);
   1826             break;
   1827         case DT_STRTAB:
   1828             si->strtab = (const char *) (si->base + *d);
   1829             break;
   1830         case DT_SYMTAB:
   1831             si->symtab = (Elf32_Sym *) (si->base + *d);
   1832             break;
   1833 #if !defined(ANDROID_SH_LINKER)
   1834         case DT_PLTREL:
   1835             if(*d != DT_REL) {
   1836                 DL_ERR("DT_RELA not supported");
   1837                 goto fail;
   1838             }
   1839             break;
   1840 #endif
   1841 #ifdef ANDROID_SH_LINKER
   1842         case DT_JMPREL:
   1843             si->plt_rela = (Elf32_Rela*) (si->base + *d);
   1844             break;
   1845         case DT_PLTRELSZ:
   1846             si->plt_rela_count = *d / sizeof(Elf32_Rela);
   1847             break;
   1848 #else
   1849         case DT_JMPREL:
   1850             si->plt_rel = (Elf32_Rel*) (si->base + *d);
   1851             break;
   1852         case DT_PLTRELSZ:
   1853             si->plt_rel_count = *d / 8;
   1854             break;
   1855 #endif
   1856         case DT_REL:
   1857             si->rel = (Elf32_Rel*) (si->base + *d);
   1858             break;
   1859         case DT_RELSZ:
   1860             si->rel_count = *d / 8;
   1861             break;
   1862 #ifdef ANDROID_SH_LINKER
   1863         case DT_RELASZ:
   1864             si->rela_count = *d / sizeof(Elf32_Rela);
   1865              break;
   1866 #endif
   1867         case DT_PLTGOT:
   1868             /* Save this in case we decide to do lazy binding. We don't yet. */
   1869             si->plt_got = (unsigned *)(si->base + *d);
   1870             break;
   1871         case DT_DEBUG:
   1872             // Set the DT_DEBUG entry to the addres of _r_debug for GDB
   1873             *d = (int) &_r_debug;
   1874             break;
   1875 #ifdef ANDROID_SH_LINKER
   1876         case DT_RELA:
   1877             si->rela = (Elf32_Rela *) (si->base + *d);
   1878             break;
   1879 #else
   1880          case DT_RELA:
   1881             DL_ERR("%5d DT_RELA not supported", pid);
   1882             goto fail;
   1883 #endif
   1884         case DT_INIT:
   1885             si->init_func = (void (*)(void))(si->base + *d);
   1886             DEBUG("%5d %s constructors (init func) found at %p\n",
   1887                   pid, si->name, si->init_func);
   1888             break;
   1889         case DT_FINI:
   1890             si->fini_func = (void (*)(void))(si->base + *d);
   1891             DEBUG("%5d %s destructors (fini func) found at %p\n",
   1892                   pid, si->name, si->fini_func);
   1893             break;
   1894         case DT_INIT_ARRAY:
   1895             si->init_array = (unsigned *)(si->base + *d);
   1896             DEBUG("%5d %s constructors (init_array) found at %p\n",
   1897                   pid, si->name, si->init_array);
   1898             break;
   1899         case DT_INIT_ARRAYSZ:
   1900             si->init_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1901             break;
   1902         case DT_FINI_ARRAY:
   1903             si->fini_array = (unsigned *)(si->base + *d);
   1904             DEBUG("%5d %s destructors (fini_array) found at %p\n",
   1905                   pid, si->name, si->fini_array);
   1906             break;
   1907         case DT_FINI_ARRAYSZ:
   1908             si->fini_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1909             break;
   1910         case DT_PREINIT_ARRAY:
   1911             si->preinit_array = (unsigned *)(si->base + *d);
   1912             DEBUG("%5d %s constructors (preinit_array) found at %p\n",
   1913                   pid, si->name, si->preinit_array);
   1914             break;
   1915         case DT_PREINIT_ARRAYSZ:
   1916             si->preinit_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1917             break;
   1918         case DT_TEXTREL:
   1919             /* TODO: make use of this. */
   1920             /* this means that we might have to write into where the text
   1921              * segment was loaded during relocation... Do something with
   1922              * it.
   1923              */
   1924             DEBUG("%5d Text segment should be writable during relocation.\n",
   1925                   pid);
   1926             break;
   1927         }
   1928     }
   1929 
   1930     DEBUG("%5d si->base = 0x%08x, si->strtab = %p, si->symtab = %p\n",
   1931            pid, si->base, si->strtab, si->symtab);
   1932 
   1933     if((si->strtab == 0) || (si->symtab == 0)) {
   1934         DL_ERR("%5d missing essential tables", pid);
   1935         goto fail;
   1936     }
   1937 
   1938     /* if this is the main executable, then load all of the preloads now */
   1939     if(si->flags & FLAG_EXE) {
   1940         int i;
   1941         memset(preloads, 0, sizeof(preloads));
   1942         for(i = 0; ldpreload_names[i] != NULL; i++) {
   1943             soinfo *lsi = find_library(ldpreload_names[i]);
   1944             if(lsi == 0) {
   1945                 strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
   1946                 DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
   1947                        pid, ldpreload_names[i], si->name, tmp_err_buf);
   1948                 goto fail;
   1949             }
   1950             lsi->refcount++;
   1951             preloads[i] = lsi;
   1952         }
   1953     }
   1954 
   1955     for(d = si->dynamic; *d; d += 2) {
   1956         if(d[0] == DT_NEEDED){
   1957             DEBUG("%5d %s needs %s\n", pid, si->name, si->strtab + d[1]);
   1958             soinfo *lsi = find_library(si->strtab + d[1]);
   1959             if(lsi == 0) {
   1960                 strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
   1961                 DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
   1962                        pid, si->strtab + d[1], si->name, tmp_err_buf);
   1963                 goto fail;
   1964             }
   1965             /* Save the soinfo of the loaded DT_NEEDED library in the payload
   1966                of the DT_NEEDED entry itself, so that we can retrieve the
   1967                soinfo directly later from the dynamic segment.  This is a hack,
   1968                but it allows us to map from DT_NEEDED to soinfo efficiently
   1969                later on when we resolve relocations, trying to look up a symgol
   1970                with dlsym().
   1971             */
   1972             d[1] = (unsigned)lsi;
   1973             lsi->refcount++;
   1974         }
   1975     }
   1976 
   1977     if(si->plt_rel) {
   1978         DEBUG("[ %5d relocating %s plt ]\n", pid, si->name );
   1979         if(reloc_library(si, si->plt_rel, si->plt_rel_count))
   1980             goto fail;
   1981     }
   1982     if(si->rel) {
   1983         DEBUG("[ %5d relocating %s ]\n", pid, si->name );
   1984         if(reloc_library(si, si->rel, si->rel_count))
   1985             goto fail;
   1986     }
   1987 
   1988 #ifdef ANDROID_SH_LINKER
   1989     if(si->plt_rela) {
   1990         DEBUG("[ %5d relocating %s plt ]\n", pid, si->name );
   1991         if(reloc_library_a(si, si->plt_rela, si->plt_rela_count))
   1992             goto fail;
   1993     }
   1994     if(si->rela) {
   1995         DEBUG("[ %5d relocating %s ]\n", pid, si->name );
   1996         if(reloc_library_a(si, si->rela, si->rela_count))
   1997             goto fail;
   1998     }
   1999 #endif /* ANDROID_SH_LINKER */
   2000 
   2001     si->flags |= FLAG_LINKED;
   2002     DEBUG("[ %5d finished linking %s ]\n", pid, si->name);
   2003 
   2004 #if 0
   2005     /* This is the way that the old dynamic linker did protection of
   2006      * non-writable areas. It would scan section headers and find where
   2007      * .text ended (rather where .data/.bss began) and assume that this is
   2008      * the upper range of the non-writable area. This is too coarse,
   2009      * and is kept here for reference until we fully move away from single
   2010      * segment elf objects. See the code in get_wr_offset (also #if'd 0)
   2011      * that made this possible.
   2012      */
   2013     if(wr_offset < 0xffffffff){
   2014         mprotect((void*) si->base, wr_offset, PROT_READ | PROT_EXEC);
   2015     }
   2016 #else
   2017     /* TODO: Verify that this does the right thing in all cases, as it
   2018      * presently probably does not. It is possible that an ELF image will
   2019      * come with multiple read-only segments. What we ought to do is scan
   2020      * the program headers again and mprotect all the read-only segments.
   2021      * To prevent re-scanning the program header, we would have to build a
   2022      * list of loadable segments in si, and then scan that instead. */
   2023     if (si->wrprotect_start != 0xffffffff && si->wrprotect_end != 0) {
   2024         mprotect((void *)si->wrprotect_start,
   2025                  si->wrprotect_end - si->wrprotect_start,
   2026                  PROT_READ | PROT_EXEC);
   2027     }
   2028 #endif
   2029 
   2030     /* If this is a SET?ID program, dup /dev/null to opened stdin,
   2031        stdout and stderr to close a security hole described in:
   2032 
   2033     ftp://ftp.freebsd.org/pub/FreeBSD/CERT/advisories/FreeBSD-SA-02:23.stdio.asc
   2034 
   2035      */
   2036     if (getuid() != geteuid() || getgid() != getegid())
   2037         nullify_closed_stdio ();
   2038     call_constructors(si);
   2039     notify_gdb_of_load(si);
   2040     return 0;
   2041 
   2042 fail:
   2043     ERROR("failed to link %s\n", si->name);
   2044     si->flags |= FLAG_ERROR;
   2045     return -1;
   2046 }
   2047 
   2048 static void parse_library_path(char *path, char *delim)
   2049 {
   2050     size_t len;
   2051     char *ldpaths_bufp = ldpaths_buf;
   2052     int i = 0;
   2053 
   2054     len = strlcpy(ldpaths_buf, path, sizeof(ldpaths_buf));
   2055 
   2056     while (i < LDPATH_MAX && (ldpaths[i] = strsep(&ldpaths_bufp, delim))) {
   2057         if (*ldpaths[i] != '\0')
   2058             ++i;
   2059     }
   2060 
   2061     /* Forget the last path if we had to truncate; this occurs if the 2nd to
   2062      * last char isn't '\0' (i.e. not originally a delim). */
   2063     if (i > 0 && len >= sizeof(ldpaths_buf) &&
   2064             ldpaths_buf[sizeof(ldpaths_buf) - 2] != '\0') {
   2065         ldpaths[i - 1] = NULL;
   2066     } else {
   2067         ldpaths[i] = NULL;
   2068     }
   2069 }
   2070 
   2071 static void parse_preloads(char *path, char *delim)
   2072 {
   2073     size_t len;
   2074     char *ldpreloads_bufp = ldpreloads_buf;
   2075     int i = 0;
   2076 
   2077     len = strlcpy(ldpreloads_buf, path, sizeof(ldpreloads_buf));
   2078 
   2079     while (i < LDPRELOAD_MAX && (ldpreload_names[i] = strsep(&ldpreloads_bufp, delim))) {
   2080         if (*ldpreload_names[i] != '\0') {
   2081             ++i;
   2082         }
   2083     }
   2084 
   2085     /* Forget the last path if we had to truncate; this occurs if the 2nd to
   2086      * last char isn't '\0' (i.e. not originally a delim). */
   2087     if (i > 0 && len >= sizeof(ldpreloads_buf) &&
   2088             ldpreloads_buf[sizeof(ldpreloads_buf) - 2] != '\0') {
   2089         ldpreload_names[i - 1] = NULL;
   2090     } else {
   2091         ldpreload_names[i] = NULL;
   2092     }
   2093 }
   2094 
   2095 int main(int argc, char **argv)
   2096 {
   2097     return 0;
   2098 }
   2099 
   2100 #define ANDROID_TLS_SLOTS  BIONIC_TLS_SLOTS
   2101 
   2102 static void * __tls_area[ANDROID_TLS_SLOTS];
   2103 
   2104 unsigned __linker_init(unsigned **elfdata)
   2105 {
   2106     static soinfo linker_soinfo;
   2107 
   2108     int argc = (int) *elfdata;
   2109     char **argv = (char**) (elfdata + 1);
   2110     unsigned *vecs = (unsigned*) (argv + argc + 1);
   2111     soinfo *si;
   2112     struct link_map * map;
   2113     char *ldpath_env = NULL;
   2114     char *ldpreload_env = NULL;
   2115 
   2116     /* Setup a temporary TLS area that is used to get a working
   2117      * errno for system calls.
   2118      */
   2119     __set_tls(__tls_area);
   2120 
   2121     pid = getpid();
   2122 
   2123 #if TIMING
   2124     struct timeval t0, t1;
   2125     gettimeofday(&t0, 0);
   2126 #endif
   2127 
   2128     /* NOTE: we store the elfdata pointer on a special location
   2129      *       of the temporary TLS area in order to pass it to
   2130      *       the C Library's runtime initializer.
   2131      *
   2132      *       The initializer must clear the slot and reset the TLS
   2133      *       to point to a different location to ensure that no other
   2134      *       shared library constructor can access it.
   2135      */
   2136     __tls_area[TLS_SLOT_BIONIC_PREINIT] = elfdata;
   2137 
   2138     debugger_init();
   2139 
   2140         /* skip past the environment */
   2141     while(vecs[0] != 0) {
   2142         if(!strncmp((char*) vecs[0], "DEBUG=", 6)) {
   2143             debug_verbosity = atoi(((char*) vecs[0]) + 6);
   2144         } else if(!strncmp((char*) vecs[0], "LD_LIBRARY_PATH=", 16)) {
   2145             ldpath_env = (char*) vecs[0] + 16;
   2146         } else if(!strncmp((char*) vecs[0], "LD_PRELOAD=", 11)) {
   2147             ldpreload_env = (char*) vecs[0] + 11;
   2148         }
   2149         vecs++;
   2150     }
   2151     vecs++;
   2152 
   2153     INFO("[ android linker & debugger ]\n");
   2154     DEBUG("%5d elfdata @ 0x%08x\n", pid, (unsigned)elfdata);
   2155 
   2156     si = alloc_info(argv[0]);
   2157     if(si == 0) {
   2158         exit(-1);
   2159     }
   2160 
   2161         /* bootstrap the link map, the main exe always needs to be first */
   2162     si->flags |= FLAG_EXE;
   2163     map = &(si->linkmap);
   2164 
   2165     map->l_addr = 0;
   2166     map->l_name = argv[0];
   2167     map->l_prev = NULL;
   2168     map->l_next = NULL;
   2169 
   2170     _r_debug.r_map = map;
   2171     r_debug_tail = map;
   2172 
   2173         /* gdb expects the linker to be in the debug shared object list,
   2174          * and we need to make sure that the reported load address is zero.
   2175          * Without this, gdb gets the wrong idea of where rtld_db_dlactivity()
   2176          * is.  Don't use alloc_info(), because the linker shouldn't
   2177          * be on the soinfo list.
   2178          */
   2179     strcpy((char*) linker_soinfo.name, "/system/bin/linker");
   2180     linker_soinfo.flags = 0;
   2181     linker_soinfo.base = 0;     // This is the important part; must be zero.
   2182     insert_soinfo_into_debug_map(&linker_soinfo);
   2183 
   2184         /* extract information passed from the kernel */
   2185     while(vecs[0] != 0){
   2186         switch(vecs[0]){
   2187         case AT_PHDR:
   2188             si->phdr = (Elf32_Phdr*) vecs[1];
   2189             break;
   2190         case AT_PHNUM:
   2191             si->phnum = (int) vecs[1];
   2192             break;
   2193         case AT_ENTRY:
   2194             si->entry = vecs[1];
   2195             break;
   2196         }
   2197         vecs += 2;
   2198     }
   2199 
   2200     ba_init(&ba_nonprelink);
   2201 
   2202     si->base = 0;
   2203     si->dynamic = (unsigned *)-1;
   2204     si->wrprotect_start = 0xffffffff;
   2205     si->wrprotect_end = 0;
   2206     si->refcount = 1;
   2207 
   2208         /* Use LD_LIBRARY_PATH if we aren't setuid/setgid */
   2209     if (ldpath_env && getuid() == geteuid() && getgid() == getegid())
   2210         parse_library_path(ldpath_env, ":");
   2211 
   2212     if (ldpreload_env && getuid() == geteuid() && getgid() == getegid()) {
   2213         parse_preloads(ldpreload_env, " :");
   2214     }
   2215 
   2216     if(link_image(si, 0)) {
   2217         char errmsg[] = "CANNOT LINK EXECUTABLE\n";
   2218         write(2, __linker_dl_err_buf, strlen(__linker_dl_err_buf));
   2219         write(2, errmsg, sizeof(errmsg));
   2220         exit(-1);
   2221     }
   2222 
   2223 #if ALLOW_SYMBOLS_FROM_MAIN
   2224     /* Set somain after we've loaded all the libraries in order to prevent
   2225      * linking of symbols back to the main image, which is not set up at that
   2226      * point yet.
   2227      */
   2228     somain = si;
   2229 #endif
   2230 
   2231 #if TIMING
   2232     gettimeofday(&t1,NULL);
   2233     PRINT("LINKER TIME: %s: %d microseconds\n", argv[0], (int) (
   2234                (((long long)t1.tv_sec * 1000000LL) + (long long)t1.tv_usec) -
   2235                (((long long)t0.tv_sec * 1000000LL) + (long long)t0.tv_usec)
   2236                ));
   2237 #endif
   2238 #if STATS
   2239     PRINT("RELO STATS: %s: %d abs, %d rel, %d copy, %d symbol\n", argv[0],
   2240            linker_stats.reloc[RELOC_ABSOLUTE],
   2241            linker_stats.reloc[RELOC_RELATIVE],
   2242            linker_stats.reloc[RELOC_COPY],
   2243            linker_stats.reloc[RELOC_SYMBOL]);
   2244 #endif
   2245 #if COUNT_PAGES
   2246     {
   2247         unsigned n;
   2248         unsigned i;
   2249         unsigned count = 0;
   2250         for(n = 0; n < 4096; n++){
   2251             if(bitmask[n]){
   2252                 unsigned x = bitmask[n];
   2253                 for(i = 0; i < 8; i++){
   2254                     if(x & 1) count++;
   2255                     x >>= 1;
   2256                 }
   2257             }
   2258         }
   2259         PRINT("PAGES MODIFIED: %s: %d (%dKB)\n", argv[0], count, count * 4);
   2260     }
   2261 #endif
   2262 
   2263 #if TIMING || STATS || COUNT_PAGES
   2264     fflush(stdout);
   2265 #endif
   2266 
   2267     TRACE("[ %5d Ready to execute '%s' @ 0x%08x ]\n", pid, si->name,
   2268           si->entry);
   2269     return si->entry;
   2270 }
   2271