Home | History | Annotate | Download | only in linker
      1 /*
      2  * Copyright (C) 2008, 2009 The Android Open Source Project
      3  * All rights reserved.
      4  *
      5  * Redistribution and use in source and binary forms, with or without
      6  * modification, are permitted provided that the following conditions
      7  * are met:
      8  *  * Redistributions of source code must retain the above copyright
      9  *    notice, this list of conditions and the following disclaimer.
     10  *  * Redistributions in binary form must reproduce the above copyright
     11  *    notice, this list of conditions and the following disclaimer in
     12  *    the documentation and/or other materials provided with the
     13  *    distribution.
     14  *
     15  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     16  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     17  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
     18  * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
     19  * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
     20  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
     21  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
     22  * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
     23  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
     24  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
     25  * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     26  * SUCH DAMAGE.
     27  */
     28 
     29 #include <linux/auxvec.h>
     30 
     31 #include <stdio.h>
     32 #include <stdlib.h>
     33 #include <string.h>
     34 #include <unistd.h>
     35 #include <fcntl.h>
     36 #include <errno.h>
     37 #include <dlfcn.h>
     38 #include <sys/stat.h>
     39 
     40 #include <pthread.h>
     41 
     42 #include <sys/mman.h>
     43 
     44 #include <sys/atomics.h>
     45 
     46 /* special private C library header - see Android.mk */
     47 #include <bionic_tls.h>
     48 
     49 #include "linker.h"
     50 #include "linker_debug.h"
     51 #include "linker_environ.h"
     52 #include "linker_format.h"
     53 
     54 #define ALLOW_SYMBOLS_FROM_MAIN 1
     55 #define SO_MAX 128
     56 
     57 /* Assume average path length of 64 and max 8 paths */
     58 #define LDPATH_BUFSIZE 512
     59 #define LDPATH_MAX 8
     60 
     61 #define LDPRELOAD_BUFSIZE 512
     62 #define LDPRELOAD_MAX 8
     63 
     64 /* >>> IMPORTANT NOTE - READ ME BEFORE MODIFYING <<<
     65  *
     66  * Do NOT use malloc() and friends or pthread_*() code here.
     67  * Don't use printf() either; it's caused mysterious memory
     68  * corruption in the past.
     69  * The linker runs before we bring up libc and it's easiest
     70  * to make sure it does not depend on any complex libc features
     71  *
     72  * open issues / todo:
     73  *
     74  * - are we doing everything we should for ARM_COPY relocations?
     75  * - cleaner error reporting
     76  * - after linking, set as much stuff as possible to READONLY
     77  *   and NOEXEC
     78  * - linker hardcodes PAGE_SIZE and PAGE_MASK because the kernel
     79  *   headers provide versions that are negative...
     80  * - allocate space for soinfo structs dynamically instead of
     81  *   having a hard limit (64)
     82 */
     83 
     84 
     85 static int link_image(soinfo *si, unsigned wr_offset);
     86 
     87 static int socount = 0;
     88 static soinfo sopool[SO_MAX];
     89 static soinfo *freelist = NULL;
     90 static soinfo *solist = &libdl_info;
     91 static soinfo *sonext = &libdl_info;
     92 #if ALLOW_SYMBOLS_FROM_MAIN
     93 static soinfo *somain; /* main process, always the one after libdl_info */
     94 #endif
     95 
     96 
     97 static inline int validate_soinfo(soinfo *si)
     98 {
     99     return (si >= sopool && si < sopool + SO_MAX) ||
    100         si == &libdl_info;
    101 }
    102 
    103 static char ldpaths_buf[LDPATH_BUFSIZE];
    104 static const char *ldpaths[LDPATH_MAX + 1];
    105 
    106 static char ldpreloads_buf[LDPRELOAD_BUFSIZE];
    107 static const char *ldpreload_names[LDPRELOAD_MAX + 1];
    108 
    109 static soinfo *preloads[LDPRELOAD_MAX + 1];
    110 
    111 int debug_verbosity;
    112 static int pid;
    113 
    114 /* This boolean is set if the program being loaded is setuid */
    115 static int program_is_setuid;
    116 
    117 #if STATS
    118 struct _link_stats linker_stats;
    119 #endif
    120 
    121 #if COUNT_PAGES
    122 unsigned bitmask[4096];
    123 #endif
    124 
    125 #ifndef PT_ARM_EXIDX
    126 #define PT_ARM_EXIDX    0x70000001      /* .ARM.exidx segment */
    127 #endif
    128 
    129 #define HOODLUM(name, ret, ...)                                               \
    130     ret name __VA_ARGS__                                                      \
    131     {                                                                         \
    132         char errstr[] = "ERROR: " #name " called from the dynamic linker!\n"; \
    133         write(2, errstr, sizeof(errstr));                                     \
    134         abort();                                                              \
    135     }
    136 HOODLUM(malloc, void *, (size_t size));
    137 HOODLUM(free, void, (void *ptr));
    138 HOODLUM(realloc, void *, (void *ptr, size_t size));
    139 HOODLUM(calloc, void *, (size_t cnt, size_t size));
    140 
    141 static char tmp_err_buf[768];
    142 static char __linker_dl_err_buf[768];
    143 #define DL_ERR(fmt, x...)                                                     \
    144     do {                                                                      \
    145         format_buffer(__linker_dl_err_buf, sizeof(__linker_dl_err_buf),            \
    146                  "%s[%d]: " fmt, __func__, __LINE__, ##x);                    \
    147         ERROR(fmt "\n", ##x);                                                      \
    148     } while(0)
    149 
    150 const char *linker_get_error(void)
    151 {
    152     return (const char *)&__linker_dl_err_buf[0];
    153 }
    154 
    155 /*
    156  * This function is an empty stub where GDB locates a breakpoint to get notified
    157  * about linker activity.
    158  */
    159 extern void __attribute__((noinline)) rtld_db_dlactivity(void);
    160 
    161 static struct r_debug _r_debug = {1, NULL, &rtld_db_dlactivity,
    162                                   RT_CONSISTENT, 0};
    163 static struct link_map *r_debug_tail = 0;
    164 
    165 static pthread_mutex_t _r_debug_lock = PTHREAD_MUTEX_INITIALIZER;
    166 
    167 static void insert_soinfo_into_debug_map(soinfo * info)
    168 {
    169     struct link_map * map;
    170 
    171     /* Copy the necessary fields into the debug structure.
    172      */
    173     map = &(info->linkmap);
    174     map->l_addr = info->base;
    175     map->l_name = (char*) info->name;
    176     map->l_ld = (uintptr_t)info->dynamic;
    177 
    178     /* Stick the new library at the end of the list.
    179      * gdb tends to care more about libc than it does
    180      * about leaf libraries, and ordering it this way
    181      * reduces the back-and-forth over the wire.
    182      */
    183     if (r_debug_tail) {
    184         r_debug_tail->l_next = map;
    185         map->l_prev = r_debug_tail;
    186         map->l_next = 0;
    187     } else {
    188         _r_debug.r_map = map;
    189         map->l_prev = 0;
    190         map->l_next = 0;
    191     }
    192     r_debug_tail = map;
    193 }
    194 
    195 static void remove_soinfo_from_debug_map(soinfo * info)
    196 {
    197     struct link_map * map = &(info->linkmap);
    198 
    199     if (r_debug_tail == map)
    200         r_debug_tail = map->l_prev;
    201 
    202     if (map->l_prev) map->l_prev->l_next = map->l_next;
    203     if (map->l_next) map->l_next->l_prev = map->l_prev;
    204 }
    205 
    206 void notify_gdb_of_load(soinfo * info)
    207 {
    208     if (info->flags & FLAG_EXE) {
    209         // GDB already knows about the main executable
    210         return;
    211     }
    212 
    213     pthread_mutex_lock(&_r_debug_lock);
    214 
    215     _r_debug.r_state = RT_ADD;
    216     rtld_db_dlactivity();
    217 
    218     insert_soinfo_into_debug_map(info);
    219 
    220     _r_debug.r_state = RT_CONSISTENT;
    221     rtld_db_dlactivity();
    222 
    223     pthread_mutex_unlock(&_r_debug_lock);
    224 }
    225 
    226 void notify_gdb_of_unload(soinfo * info)
    227 {
    228     if (info->flags & FLAG_EXE) {
    229         // GDB already knows about the main executable
    230         return;
    231     }
    232 
    233     pthread_mutex_lock(&_r_debug_lock);
    234 
    235     _r_debug.r_state = RT_DELETE;
    236     rtld_db_dlactivity();
    237 
    238     remove_soinfo_from_debug_map(info);
    239 
    240     _r_debug.r_state = RT_CONSISTENT;
    241     rtld_db_dlactivity();
    242 
    243     pthread_mutex_unlock(&_r_debug_lock);
    244 }
    245 
    246 void notify_gdb_of_libraries()
    247 {
    248     _r_debug.r_state = RT_ADD;
    249     rtld_db_dlactivity();
    250     _r_debug.r_state = RT_CONSISTENT;
    251     rtld_db_dlactivity();
    252 }
    253 
    254 static soinfo *alloc_info(const char *name)
    255 {
    256     soinfo *si;
    257 
    258     if(strlen(name) >= SOINFO_NAME_LEN) {
    259         DL_ERR("%5d library name %s too long", pid, name);
    260         return NULL;
    261     }
    262 
    263     /* The freelist is populated when we call free_info(), which in turn is
    264        done only by dlclose(), which is not likely to be used.
    265     */
    266     if (!freelist) {
    267         if(socount == SO_MAX) {
    268             DL_ERR("%5d too many libraries when loading %s", pid, name);
    269             return NULL;
    270         }
    271         freelist = sopool + socount++;
    272         freelist->next = NULL;
    273     }
    274 
    275     si = freelist;
    276     freelist = freelist->next;
    277 
    278     /* Make sure we get a clean block of soinfo */
    279     memset(si, 0, sizeof(soinfo));
    280     strlcpy((char*) si->name, name, sizeof(si->name));
    281     sonext->next = si;
    282     si->next = NULL;
    283     si->refcount = 0;
    284     sonext = si;
    285 
    286     TRACE("%5d name %s: allocated soinfo @ %p\n", pid, name, si);
    287     return si;
    288 }
    289 
    290 static void free_info(soinfo *si)
    291 {
    292     soinfo *prev = NULL, *trav;
    293 
    294     TRACE("%5d name %s: freeing soinfo @ %p\n", pid, si->name, si);
    295 
    296     for(trav = solist; trav != NULL; trav = trav->next){
    297         if (trav == si)
    298             break;
    299         prev = trav;
    300     }
    301     if (trav == NULL) {
    302         /* si was not ni solist */
    303         DL_ERR("%5d name %s is not in solist!", pid, si->name);
    304         return;
    305     }
    306 
    307     /* prev will never be NULL, because the first entry in solist is
    308        always the static libdl_info.
    309     */
    310     prev->next = si->next;
    311     if (si == sonext) sonext = prev;
    312     si->next = freelist;
    313     freelist = si;
    314 }
    315 
    316 #ifndef LINKER_TEXT_BASE
    317 #error "linker's makefile must define LINKER_TEXT_BASE"
    318 #endif
    319 #ifndef LINKER_AREA_SIZE
    320 #error "linker's makefile must define LINKER_AREA_SIZE"
    321 #endif
    322 #define LINKER_BASE ((LINKER_TEXT_BASE) & 0xfff00000)
    323 #define LINKER_TOP  (LINKER_BASE + (LINKER_AREA_SIZE))
    324 
    325 const char *addr_to_name(unsigned addr)
    326 {
    327     soinfo *si;
    328 
    329     for(si = solist; si != 0; si = si->next){
    330         if((addr >= si->base) && (addr < (si->base + si->size))) {
    331             return si->name;
    332         }
    333     }
    334 
    335     if((addr >= LINKER_BASE) && (addr < LINKER_TOP)){
    336         return "linker";
    337     }
    338 
    339     return "";
    340 }
    341 
    342 /* For a given PC, find the .so that it belongs to.
    343  * Returns the base address of the .ARM.exidx section
    344  * for that .so, and the number of 8-byte entries
    345  * in that section (via *pcount).
    346  *
    347  * Intended to be called by libc's __gnu_Unwind_Find_exidx().
    348  *
    349  * This function is exposed via dlfcn.c and libdl.so.
    350  */
    351 #ifdef ANDROID_ARM_LINKER
    352 _Unwind_Ptr dl_unwind_find_exidx(_Unwind_Ptr pc, int *pcount)
    353 {
    354     soinfo *si;
    355     unsigned addr = (unsigned)pc;
    356 
    357     if ((addr < LINKER_BASE) || (addr >= LINKER_TOP)) {
    358         for (si = solist; si != 0; si = si->next){
    359             if ((addr >= si->base) && (addr < (si->base + si->size))) {
    360                 *pcount = si->ARM_exidx_count;
    361                 return (_Unwind_Ptr)(si->base + (unsigned long)si->ARM_exidx);
    362             }
    363         }
    364     }
    365    *pcount = 0;
    366     return NULL;
    367 }
    368 #elif defined(ANDROID_X86_LINKER) || defined(ANDROID_SH_LINKER)
    369 /* Here, we only have to provide a callback to iterate across all the
    370  * loaded libraries. gcc_eh does the rest. */
    371 int
    372 dl_iterate_phdr(int (*cb)(struct dl_phdr_info *info, size_t size, void *data),
    373                 void *data)
    374 {
    375     soinfo *si;
    376     struct dl_phdr_info dl_info;
    377     int rv = 0;
    378 
    379     for (si = solist; si != NULL; si = si->next) {
    380         dl_info.dlpi_addr = si->linkmap.l_addr;
    381         dl_info.dlpi_name = si->linkmap.l_name;
    382         dl_info.dlpi_phdr = si->phdr;
    383         dl_info.dlpi_phnum = si->phnum;
    384         rv = cb(&dl_info, sizeof (struct dl_phdr_info), data);
    385         if (rv != 0)
    386             break;
    387     }
    388     return rv;
    389 }
    390 #endif
    391 
    392 static Elf32_Sym *_elf_lookup(soinfo *si, unsigned hash, const char *name)
    393 {
    394     Elf32_Sym *s;
    395     Elf32_Sym *symtab = si->symtab;
    396     const char *strtab = si->strtab;
    397     unsigned n;
    398 
    399     TRACE_TYPE(LOOKUP, "%5d SEARCH %s in %s@0x%08x %08x %d\n", pid,
    400                name, si->name, si->base, hash, hash % si->nbucket);
    401     n = hash % si->nbucket;
    402 
    403     for(n = si->bucket[hash % si->nbucket]; n != 0; n = si->chain[n]){
    404         s = symtab + n;
    405         if(strcmp(strtab + s->st_name, name)) continue;
    406 
    407             /* only concern ourselves with global and weak symbol definitions */
    408         switch(ELF32_ST_BIND(s->st_info)){
    409         case STB_GLOBAL:
    410         case STB_WEAK:
    411                 /* no section == undefined */
    412             if(s->st_shndx == 0) continue;
    413 
    414             TRACE_TYPE(LOOKUP, "%5d FOUND %s in %s (%08x) %d\n", pid,
    415                        name, si->name, s->st_value, s->st_size);
    416             return s;
    417         }
    418     }
    419 
    420     return NULL;
    421 }
    422 
    423 static unsigned elfhash(const char *_name)
    424 {
    425     const unsigned char *name = (const unsigned char *) _name;
    426     unsigned h = 0, g;
    427 
    428     while(*name) {
    429         h = (h << 4) + *name++;
    430         g = h & 0xf0000000;
    431         h ^= g;
    432         h ^= g >> 24;
    433     }
    434     return h;
    435 }
    436 
    437 static Elf32_Sym *
    438 _do_lookup(soinfo *si, const char *name, unsigned *base)
    439 {
    440     unsigned elf_hash = elfhash(name);
    441     Elf32_Sym *s;
    442     unsigned *d;
    443     soinfo *lsi = si;
    444     int i;
    445 
    446     /* Look for symbols in the local scope first (the object who is
    447      * searching). This happens with C++ templates on i386 for some
    448      * reason.
    449      *
    450      * Notes on weak symbols:
    451      * The ELF specs are ambigious about treatment of weak definitions in
    452      * dynamic linking.  Some systems return the first definition found
    453      * and some the first non-weak definition.   This is system dependent.
    454      * Here we return the first definition found for simplicity.  */
    455     s = _elf_lookup(si, elf_hash, name);
    456     if(s != NULL)
    457         goto done;
    458 
    459     /* Next, look for it in the preloads list */
    460     for(i = 0; preloads[i] != NULL; i++) {
    461         lsi = preloads[i];
    462         s = _elf_lookup(lsi, elf_hash, name);
    463         if(s != NULL)
    464             goto done;
    465     }
    466 
    467     for(d = si->dynamic; *d; d += 2) {
    468         if(d[0] == DT_NEEDED){
    469             lsi = (soinfo *)d[1];
    470             if (!validate_soinfo(lsi)) {
    471                 DL_ERR("%5d bad DT_NEEDED pointer in %s",
    472                        pid, si->name);
    473                 return NULL;
    474             }
    475 
    476             DEBUG("%5d %s: looking up %s in %s\n",
    477                   pid, si->name, name, lsi->name);
    478             s = _elf_lookup(lsi, elf_hash, name);
    479             if ((s != NULL) && (s->st_shndx != SHN_UNDEF))
    480                 goto done;
    481         }
    482     }
    483 
    484 #if ALLOW_SYMBOLS_FROM_MAIN
    485     /* If we are resolving relocations while dlopen()ing a library, it's OK for
    486      * the library to resolve a symbol that's defined in the executable itself,
    487      * although this is rare and is generally a bad idea.
    488      */
    489     if (somain) {
    490         lsi = somain;
    491         DEBUG("%5d %s: looking up %s in executable %s\n",
    492               pid, si->name, name, lsi->name);
    493         s = _elf_lookup(lsi, elf_hash, name);
    494     }
    495 #endif
    496 
    497 done:
    498     if(s != NULL) {
    499         TRACE_TYPE(LOOKUP, "%5d si %s sym %s s->st_value = 0x%08x, "
    500                    "found in %s, base = 0x%08x\n",
    501                    pid, si->name, name, s->st_value, lsi->name, lsi->base);
    502         *base = lsi->base;
    503         return s;
    504     }
    505 
    506     return NULL;
    507 }
    508 
    509 /* This is used by dl_sym().  It performs symbol lookup only within the
    510    specified soinfo object and not in any of its dependencies.
    511  */
    512 Elf32_Sym *lookup_in_library(soinfo *si, const char *name)
    513 {
    514     return _elf_lookup(si, elfhash(name), name);
    515 }
    516 
    517 /* This is used by dl_sym().  It performs a global symbol lookup.
    518  */
    519 Elf32_Sym *lookup(const char *name, soinfo **found, soinfo *start)
    520 {
    521     unsigned elf_hash = elfhash(name);
    522     Elf32_Sym *s = NULL;
    523     soinfo *si;
    524 
    525     if(start == NULL) {
    526         start = solist;
    527     }
    528 
    529     for(si = start; (s == NULL) && (si != NULL); si = si->next)
    530     {
    531         if(si->flags & FLAG_ERROR)
    532             continue;
    533         s = _elf_lookup(si, elf_hash, name);
    534         if (s != NULL) {
    535             *found = si;
    536             break;
    537         }
    538     }
    539 
    540     if(s != NULL) {
    541         TRACE_TYPE(LOOKUP, "%5d %s s->st_value = 0x%08x, "
    542                    "si->base = 0x%08x\n", pid, name, s->st_value, si->base);
    543         return s;
    544     }
    545 
    546     return NULL;
    547 }
    548 
    549 soinfo *find_containing_library(const void *addr)
    550 {
    551     soinfo *si;
    552 
    553     for(si = solist; si != NULL; si = si->next)
    554     {
    555         if((unsigned)addr >= si->base && (unsigned)addr - si->base < si->size) {
    556             return si;
    557         }
    558     }
    559 
    560     return NULL;
    561 }
    562 
    563 Elf32_Sym *find_containing_symbol(const void *addr, soinfo *si)
    564 {
    565     unsigned int i;
    566     unsigned soaddr = (unsigned)addr - si->base;
    567 
    568     /* Search the library's symbol table for any defined symbol which
    569      * contains this address */
    570     for(i=0; i<si->nchain; i++) {
    571         Elf32_Sym *sym = &si->symtab[i];
    572 
    573         if(sym->st_shndx != SHN_UNDEF &&
    574            soaddr >= sym->st_value &&
    575            soaddr < sym->st_value + sym->st_size) {
    576             return sym;
    577         }
    578     }
    579 
    580     return NULL;
    581 }
    582 
    583 #if 0
    584 static void dump(soinfo *si)
    585 {
    586     Elf32_Sym *s = si->symtab;
    587     unsigned n;
    588 
    589     for(n = 0; n < si->nchain; n++) {
    590         TRACE("%5d %04d> %08x: %02x %04x %08x %08x %s\n", pid, n, s,
    591                s->st_info, s->st_shndx, s->st_value, s->st_size,
    592                si->strtab + s->st_name);
    593         s++;
    594     }
    595 }
    596 #endif
    597 
    598 static const char *sopaths[] = {
    599     "/vendor/lib",
    600     "/system/lib",
    601     0
    602 };
    603 
    604 static int _open_lib(const char *name)
    605 {
    606     int fd;
    607     struct stat filestat;
    608 
    609     if ((stat(name, &filestat) >= 0) && S_ISREG(filestat.st_mode)) {
    610         if ((fd = open(name, O_RDONLY)) >= 0)
    611             return fd;
    612     }
    613 
    614     return -1;
    615 }
    616 
    617 static int open_library(const char *name)
    618 {
    619     int fd;
    620     char buf[512];
    621     const char **path;
    622     int n;
    623 
    624     TRACE("[ %5d opening %s ]\n", pid, name);
    625 
    626     if(name == 0) return -1;
    627     if(strlen(name) > 256) return -1;
    628 
    629     if ((name[0] == '/') && ((fd = _open_lib(name)) >= 0))
    630         return fd;
    631 
    632     for (path = ldpaths; *path; path++) {
    633         n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
    634         if (n < 0 || n >= (int)sizeof(buf)) {
    635             WARN("Ignoring very long library path: %s/%s\n", *path, name);
    636             continue;
    637         }
    638         if ((fd = _open_lib(buf)) >= 0)
    639             return fd;
    640     }
    641     for (path = sopaths; *path; path++) {
    642         n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
    643         if (n < 0 || n >= (int)sizeof(buf)) {
    644             WARN("Ignoring very long library path: %s/%s\n", *path, name);
    645             continue;
    646         }
    647         if ((fd = _open_lib(buf)) >= 0)
    648             return fd;
    649     }
    650 
    651     return -1;
    652 }
    653 
    654 /* temporary space for holding the first page of the shared lib
    655  * which contains the elf header (with the pht). */
    656 static unsigned char __header[PAGE_SIZE];
    657 
    658 typedef struct {
    659     long mmap_addr;
    660     char tag[4]; /* 'P', 'R', 'E', ' ' */
    661 } prelink_info_t;
    662 
    663 /* Returns the requested base address if the library is prelinked,
    664  * and 0 otherwise.  */
    665 static unsigned long
    666 is_prelinked(int fd, const char *name)
    667 {
    668     off_t sz;
    669     prelink_info_t info;
    670 
    671     sz = lseek(fd, -sizeof(prelink_info_t), SEEK_END);
    672     if (sz < 0) {
    673         DL_ERR("lseek() failed!");
    674         return 0;
    675     }
    676 
    677     if (read(fd, &info, sizeof(info)) != sizeof(info)) {
    678         WARN("Could not read prelink_info_t structure for `%s`\n", name);
    679         return 0;
    680     }
    681 
    682     if (strncmp(info.tag, "PRE ", 4)) {
    683         WARN("`%s` is not a prelinked library\n", name);
    684         return 0;
    685     }
    686 
    687     return (unsigned long)info.mmap_addr;
    688 }
    689 
    690 /* verify_elf_object
    691  *      Verifies if the object @ base is a valid ELF object
    692  *
    693  * Args:
    694  *
    695  * Returns:
    696  *       0 on success
    697  *      -1 if no valid ELF object is found @ base.
    698  */
    699 static int
    700 verify_elf_object(void *base, const char *name)
    701 {
    702     Elf32_Ehdr *hdr = (Elf32_Ehdr *) base;
    703 
    704     if (hdr->e_ident[EI_MAG0] != ELFMAG0) return -1;
    705     if (hdr->e_ident[EI_MAG1] != ELFMAG1) return -1;
    706     if (hdr->e_ident[EI_MAG2] != ELFMAG2) return -1;
    707     if (hdr->e_ident[EI_MAG3] != ELFMAG3) return -1;
    708 
    709     /* TODO: Should we verify anything else in the header? */
    710 
    711     return 0;
    712 }
    713 
    714 
    715 /* get_lib_extents
    716  *      Retrieves the base (*base) address where the ELF object should be
    717  *      mapped and its overall memory size (*total_sz).
    718  *
    719  * Args:
    720  *      fd: Opened file descriptor for the library
    721  *      name: The name of the library
    722  *      _hdr: Pointer to the header page of the library
    723  *      total_sz: Total size of the memory that should be allocated for
    724  *                this library
    725  *
    726  * Returns:
    727  *      -1 if there was an error while trying to get the lib extents.
    728  *         The possible reasons are:
    729  *             - Could not determine if the library was prelinked.
    730  *             - The library provided is not a valid ELF object
    731  *       0 if the library did not request a specific base offset (normal
    732  *         for non-prelinked libs)
    733  *     > 0 if the library requests a specific address to be mapped to.
    734  *         This indicates a pre-linked library.
    735  */
    736 static unsigned
    737 get_lib_extents(int fd, const char *name, void *__hdr, unsigned *total_sz)
    738 {
    739     unsigned req_base;
    740     unsigned min_vaddr = 0xffffffff;
    741     unsigned max_vaddr = 0;
    742     unsigned char *_hdr = (unsigned char *)__hdr;
    743     Elf32_Ehdr *ehdr = (Elf32_Ehdr *)_hdr;
    744     Elf32_Phdr *phdr;
    745     int cnt;
    746 
    747     TRACE("[ %5d Computing extents for '%s'. ]\n", pid, name);
    748     if (verify_elf_object(_hdr, name) < 0) {
    749         DL_ERR("%5d - %s is not a valid ELF object", pid, name);
    750         return (unsigned)-1;
    751     }
    752 
    753     req_base = (unsigned) is_prelinked(fd, name);
    754     if (req_base == (unsigned)-1)
    755         return -1;
    756     else if (req_base != 0) {
    757         TRACE("[ %5d - Prelinked library '%s' requesting base @ 0x%08x ]\n",
    758               pid, name, req_base);
    759     } else {
    760         TRACE("[ %5d - Non-prelinked library '%s' found. ]\n", pid, name);
    761     }
    762 
    763     phdr = (Elf32_Phdr *)(_hdr + ehdr->e_phoff);
    764 
    765     /* find the min/max p_vaddrs from all the PT_LOAD segments so we can
    766      * get the range. */
    767     for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
    768         if (phdr->p_type == PT_LOAD) {
    769             if ((phdr->p_vaddr + phdr->p_memsz) > max_vaddr)
    770                 max_vaddr = phdr->p_vaddr + phdr->p_memsz;
    771             if (phdr->p_vaddr < min_vaddr)
    772                 min_vaddr = phdr->p_vaddr;
    773         }
    774     }
    775 
    776     if ((min_vaddr == 0xffffffff) && (max_vaddr == 0)) {
    777         DL_ERR("%5d - No loadable segments found in %s.", pid, name);
    778         return (unsigned)-1;
    779     }
    780 
    781     /* truncate min_vaddr down to page boundary */
    782     min_vaddr &= ~PAGE_MASK;
    783 
    784     /* round max_vaddr up to the next page */
    785     max_vaddr = (max_vaddr + PAGE_SIZE - 1) & ~PAGE_MASK;
    786 
    787     *total_sz = (max_vaddr - min_vaddr);
    788     return (unsigned)req_base;
    789 }
    790 
    791 /* alloc_mem_region
    792  *
    793  *     This function reserves a chunk of memory to be used for mapping in
    794  *     the shared library. We reserve the entire memory region here, and
    795  *     then the rest of the linker will relocate the individual loadable
    796  *     segments into the correct locations within this memory range.
    797  *
    798  * Args:
    799  *     si->base: The requested base of the allocation. If 0, a sane one will be
    800  *               chosen in the range LIBBASE <= base < LIBLAST.
    801  *     si->size: The size of the allocation.
    802  *
    803  * Returns:
    804  *     -1 on failure, and 0 on success.  On success, si->base will contain
    805  *     the virtual address at which the library will be mapped.
    806  */
    807 
    808 static int reserve_mem_region(soinfo *si)
    809 {
    810     void *base = mmap((void *)si->base, si->size, PROT_READ | PROT_EXEC,
    811                       MAP_FIXED | MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
    812     if (base == MAP_FAILED) {
    813         DL_ERR("%5d can NOT map (%sprelinked) library '%s' at 0x%08x "
    814               "as requested, will try general pool: %d (%s)",
    815               pid, (si->base ? "" : "non-"), si->name, si->base,
    816               errno, strerror(errno));
    817         return -1;
    818     } else if (base != (void *)si->base) {
    819         DL_ERR("OOPS: %5d %sprelinked library '%s' mapped at 0x%08x, "
    820               "not at 0x%08x", pid, (si->base ? "" : "non-"),
    821               si->name, (unsigned)base, si->base);
    822         munmap(base, si->size);
    823         return -1;
    824     }
    825     return 0;
    826 }
    827 
    828 static int
    829 alloc_mem_region(soinfo *si)
    830 {
    831     if (si->base) {
    832         /* Attempt to mmap a prelinked library. */
    833         return reserve_mem_region(si);
    834     }
    835 
    836     /* This is not a prelinked library, so we use the kernel's default
    837        allocator.
    838     */
    839 
    840     void *base = mmap(NULL, si->size, PROT_READ | PROT_EXEC,
    841                       MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
    842     if (base == MAP_FAILED) {
    843         DL_ERR("%5d mmap of library '%s' failed: %d (%s)\n",
    844               pid, si->name,
    845               errno, strerror(errno));
    846         goto err;
    847     }
    848     si->base = (unsigned) base;
    849     PRINT("%5d mapped library '%s' to %08x via kernel allocator.\n",
    850           pid, si->name, si->base);
    851     return 0;
    852 
    853 err:
    854     DL_ERR("OOPS: %5d cannot map library '%s'. no vspace available.",
    855           pid, si->name);
    856     return -1;
    857 }
    858 
    859 #define MAYBE_MAP_FLAG(x,from,to)    (((x) & (from)) ? (to) : 0)
    860 #define PFLAGS_TO_PROT(x)            (MAYBE_MAP_FLAG((x), PF_X, PROT_EXEC) | \
    861                                       MAYBE_MAP_FLAG((x), PF_R, PROT_READ) | \
    862                                       MAYBE_MAP_FLAG((x), PF_W, PROT_WRITE))
    863 /* load_segments
    864  *
    865  *     This function loads all the loadable (PT_LOAD) segments into memory
    866  *     at their appropriate memory offsets off the base address.
    867  *
    868  * Args:
    869  *     fd: Open file descriptor to the library to load.
    870  *     header: Pointer to a header page that contains the ELF header.
    871  *             This is needed since we haven't mapped in the real file yet.
    872  *     si: ptr to soinfo struct describing the shared object.
    873  *
    874  * Returns:
    875  *     0 on success, -1 on failure.
    876  */
    877 static int
    878 load_segments(int fd, void *header, soinfo *si)
    879 {
    880     Elf32_Ehdr *ehdr = (Elf32_Ehdr *)header;
    881     Elf32_Phdr *phdr = (Elf32_Phdr *)((unsigned char *)header + ehdr->e_phoff);
    882     unsigned char *base = (unsigned char *)si->base;
    883     int cnt;
    884     unsigned len;
    885     unsigned char *tmp;
    886     unsigned char *pbase;
    887     unsigned char *extra_base;
    888     unsigned extra_len;
    889     unsigned total_sz = 0;
    890 
    891     si->wrprotect_start = 0xffffffff;
    892     si->wrprotect_end = 0;
    893 
    894     TRACE("[ %5d - Begin loading segments for '%s' @ 0x%08x ]\n",
    895           pid, si->name, (unsigned)si->base);
    896     /* Now go through all the PT_LOAD segments and map them into memory
    897      * at the appropriate locations. */
    898     for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
    899         if (phdr->p_type == PT_LOAD) {
    900             DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
    901             /* we want to map in the segment on a page boundary */
    902             tmp = base + (phdr->p_vaddr & (~PAGE_MASK));
    903             /* add the # of bytes we masked off above to the total length. */
    904             len = phdr->p_filesz + (phdr->p_vaddr & PAGE_MASK);
    905 
    906             TRACE("[ %d - Trying to load segment from '%s' @ 0x%08x "
    907                   "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x ]\n", pid, si->name,
    908                   (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
    909             pbase = mmap(tmp, len, PFLAGS_TO_PROT(phdr->p_flags),
    910                          MAP_PRIVATE | MAP_FIXED, fd,
    911                          phdr->p_offset & (~PAGE_MASK));
    912             if (pbase == MAP_FAILED) {
    913                 DL_ERR("%d failed to map segment from '%s' @ 0x%08x (0x%08x). "
    914                       "p_vaddr=0x%08x p_offset=0x%08x", pid, si->name,
    915                       (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
    916                 goto fail;
    917             }
    918 
    919             /* If 'len' didn't end on page boundary, and it's a writable
    920              * segment, zero-fill the rest. */
    921             if ((len & PAGE_MASK) && (phdr->p_flags & PF_W))
    922                 memset((void *)(pbase + len), 0, PAGE_SIZE - (len & PAGE_MASK));
    923 
    924             /* Check to see if we need to extend the map for this segment to
    925              * cover the diff between filesz and memsz (i.e. for bss).
    926              *
    927              *  base           _+---------------------+  page boundary
    928              *                  .                     .
    929              *                  |                     |
    930              *                  .                     .
    931              *  pbase          _+---------------------+  page boundary
    932              *                  |                     |
    933              *                  .                     .
    934              *  base + p_vaddr _|                     |
    935              *                  . \          \        .
    936              *                  . | filesz   |        .
    937              *  pbase + len    _| /          |        |
    938              *     <0 pad>      .            .        .
    939              *  extra_base     _+------------|--------+  page boundary
    940              *               /  .            .        .
    941              *               |  .            .        .
    942              *               |  +------------|--------+  page boundary
    943              *  extra_len->  |  |            |        |
    944              *               |  .            | memsz  .
    945              *               |  .            |        .
    946              *               \ _|            /        |
    947              *                  .                     .
    948              *                  |                     |
    949              *                 _+---------------------+  page boundary
    950              */
    951             tmp = (unsigned char *)(((unsigned)pbase + len + PAGE_SIZE - 1) &
    952                                     (~PAGE_MASK));
    953             if (tmp < (base + phdr->p_vaddr + phdr->p_memsz)) {
    954                 extra_len = base + phdr->p_vaddr + phdr->p_memsz - tmp;
    955                 TRACE("[ %5d - Need to extend segment from '%s' @ 0x%08x "
    956                       "(0x%08x) ]\n", pid, si->name, (unsigned)tmp, extra_len);
    957                 /* map in the extra page(s) as anonymous into the range.
    958                  * This is probably not necessary as we already mapped in
    959                  * the entire region previously, but we just want to be
    960                  * sure. This will also set the right flags on the region
    961                  * (though we can probably accomplish the same thing with
    962                  * mprotect).
    963                  */
    964                 extra_base = mmap((void *)tmp, extra_len,
    965                                   PFLAGS_TO_PROT(phdr->p_flags),
    966                                   MAP_PRIVATE | MAP_FIXED | MAP_ANONYMOUS,
    967                                   -1, 0);
    968                 if (extra_base == MAP_FAILED) {
    969                     DL_ERR("[ %5d - failed to extend segment from '%s' @ 0x%08x"
    970                            " (0x%08x) ]", pid, si->name, (unsigned)tmp,
    971                           extra_len);
    972                     goto fail;
    973                 }
    974                 /* TODO: Check if we need to memset-0 this region.
    975                  * Anonymous mappings are zero-filled copy-on-writes, so we
    976                  * shouldn't need to. */
    977                 TRACE("[ %5d - Segment from '%s' extended @ 0x%08x "
    978                       "(0x%08x)\n", pid, si->name, (unsigned)extra_base,
    979                       extra_len);
    980             }
    981             /* set the len here to show the full extent of the segment we
    982              * just loaded, mostly for debugging */
    983             len = (((unsigned)base + phdr->p_vaddr + phdr->p_memsz +
    984                     PAGE_SIZE - 1) & (~PAGE_MASK)) - (unsigned)pbase;
    985             TRACE("[ %5d - Successfully loaded segment from '%s' @ 0x%08x "
    986                   "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x\n", pid, si->name,
    987                   (unsigned)pbase, len, phdr->p_vaddr, phdr->p_offset);
    988             total_sz += len;
    989             /* Make the section writable just in case we'll have to write to
    990              * it during relocation (i.e. text segment). However, we will
    991              * remember what range of addresses should be write protected.
    992              *
    993              */
    994             if (!(phdr->p_flags & PF_W)) {
    995                 if ((unsigned)pbase < si->wrprotect_start)
    996                     si->wrprotect_start = (unsigned)pbase;
    997                 if (((unsigned)pbase + len) > si->wrprotect_end)
    998                     si->wrprotect_end = (unsigned)pbase + len;
    999                 mprotect(pbase, len,
   1000                          PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
   1001             }
   1002         } else if (phdr->p_type == PT_DYNAMIC) {
   1003             DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
   1004             /* this segment contains the dynamic linking information */
   1005             si->dynamic = (unsigned *)(base + phdr->p_vaddr);
   1006         } else {
   1007 #ifdef ANDROID_ARM_LINKER
   1008             if (phdr->p_type == PT_ARM_EXIDX) {
   1009                 DEBUG_DUMP_PHDR(phdr, "PT_ARM_EXIDX", pid);
   1010                 /* exidx entries (used for stack unwinding) are 8 bytes each.
   1011                  */
   1012                 si->ARM_exidx = (unsigned *)phdr->p_vaddr;
   1013                 si->ARM_exidx_count = phdr->p_memsz / 8;
   1014             }
   1015 #endif
   1016         }
   1017 
   1018     }
   1019 
   1020     /* Sanity check */
   1021     if (total_sz > si->size) {
   1022         DL_ERR("%5d - Total length (0x%08x) of mapped segments from '%s' is "
   1023               "greater than what was allocated (0x%08x). THIS IS BAD!",
   1024               pid, total_sz, si->name, si->size);
   1025         goto fail;
   1026     }
   1027 
   1028     TRACE("[ %5d - Finish loading segments for '%s' @ 0x%08x. "
   1029           "Total memory footprint: 0x%08x bytes ]\n", pid, si->name,
   1030           (unsigned)si->base, si->size);
   1031     return 0;
   1032 
   1033 fail:
   1034     /* We can just blindly unmap the entire region even though some things
   1035      * were mapped in originally with anonymous and others could have been
   1036      * been mapped in from the file before we failed. The kernel will unmap
   1037      * all the pages in the range, irrespective of how they got there.
   1038      */
   1039     munmap((void *)si->base, si->size);
   1040     si->flags |= FLAG_ERROR;
   1041     return -1;
   1042 }
   1043 
   1044 /* TODO: Implement this to take care of the fact that Android ARM
   1045  * ELF objects shove everything into a single loadable segment that has the
   1046  * write bit set. wr_offset is then used to set non-(data|bss) pages to be
   1047  * non-writable.
   1048  */
   1049 #if 0
   1050 static unsigned
   1051 get_wr_offset(int fd, const char *name, Elf32_Ehdr *ehdr)
   1052 {
   1053     Elf32_Shdr *shdr_start;
   1054     Elf32_Shdr *shdr;
   1055     int shdr_sz = ehdr->e_shnum * sizeof(Elf32_Shdr);
   1056     int cnt;
   1057     unsigned wr_offset = 0xffffffff;
   1058 
   1059     shdr_start = mmap(0, shdr_sz, PROT_READ, MAP_PRIVATE, fd,
   1060                       ehdr->e_shoff & (~PAGE_MASK));
   1061     if (shdr_start == MAP_FAILED) {
   1062         WARN("%5d - Could not read section header info from '%s'. Will not "
   1063              "not be able to determine write-protect offset.\n", pid, name);
   1064         return (unsigned)-1;
   1065     }
   1066 
   1067     for(cnt = 0, shdr = shdr_start; cnt < ehdr->e_shnum; ++cnt, ++shdr) {
   1068         if ((shdr->sh_type != SHT_NULL) && (shdr->sh_flags & SHF_WRITE) &&
   1069             (shdr->sh_addr < wr_offset)) {
   1070             wr_offset = shdr->sh_addr;
   1071         }
   1072     }
   1073 
   1074     munmap(shdr_start, shdr_sz);
   1075     return wr_offset;
   1076 }
   1077 #endif
   1078 
   1079 static soinfo *
   1080 load_library(const char *name)
   1081 {
   1082     int fd = open_library(name);
   1083     int cnt;
   1084     unsigned ext_sz;
   1085     unsigned req_base;
   1086     const char *bname;
   1087     soinfo *si = NULL;
   1088     Elf32_Ehdr *hdr;
   1089 
   1090     if(fd == -1) {
   1091         DL_ERR("Library '%s' not found", name);
   1092         return NULL;
   1093     }
   1094 
   1095     /* We have to read the ELF header to figure out what to do with this image
   1096      */
   1097     if (lseek(fd, 0, SEEK_SET) < 0) {
   1098         DL_ERR("lseek() failed!");
   1099         goto fail;
   1100     }
   1101 
   1102     if ((cnt = read(fd, &__header[0], PAGE_SIZE)) < 0) {
   1103         DL_ERR("read() failed!");
   1104         goto fail;
   1105     }
   1106 
   1107     /* Parse the ELF header and get the size of the memory footprint for
   1108      * the library */
   1109     req_base = get_lib_extents(fd, name, &__header[0], &ext_sz);
   1110     if (req_base == (unsigned)-1)
   1111         goto fail;
   1112     TRACE("[ %5d - '%s' (%s) wants base=0x%08x sz=0x%08x ]\n", pid, name,
   1113           (req_base ? "prelinked" : "not pre-linked"), req_base, ext_sz);
   1114 
   1115     /* Now configure the soinfo struct where we'll store all of our data
   1116      * for the ELF object. If the loading fails, we waste the entry, but
   1117      * same thing would happen if we failed during linking. Configuring the
   1118      * soinfo struct here is a lot more convenient.
   1119      */
   1120     bname = strrchr(name, '/');
   1121     si = alloc_info(bname ? bname + 1 : name);
   1122     if (si == NULL)
   1123         goto fail;
   1124 
   1125     /* Carve out a chunk of memory where we will map in the individual
   1126      * segments */
   1127     si->base = req_base;
   1128     si->size = ext_sz;
   1129     si->flags = 0;
   1130     si->entry = 0;
   1131     si->dynamic = (unsigned *)-1;
   1132     if (alloc_mem_region(si) < 0)
   1133         goto fail;
   1134 
   1135     TRACE("[ %5d allocated memory for %s @ %p (0x%08x) ]\n",
   1136           pid, name, (void *)si->base, (unsigned) ext_sz);
   1137 
   1138     /* Now actually load the library's segments into right places in memory */
   1139     if (load_segments(fd, &__header[0], si) < 0) {
   1140         goto fail;
   1141     }
   1142 
   1143     /* this might not be right. Technically, we don't even need this info
   1144      * once we go through 'load_segments'. */
   1145     hdr = (Elf32_Ehdr *)si->base;
   1146     si->phdr = (Elf32_Phdr *)((unsigned char *)si->base + hdr->e_phoff);
   1147     si->phnum = hdr->e_phnum;
   1148     /**/
   1149 
   1150     close(fd);
   1151     return si;
   1152 
   1153 fail:
   1154     if (si) free_info(si);
   1155     close(fd);
   1156     return NULL;
   1157 }
   1158 
   1159 static soinfo *
   1160 init_library(soinfo *si)
   1161 {
   1162     unsigned wr_offset = 0xffffffff;
   1163 
   1164     /* At this point we know that whatever is loaded @ base is a valid ELF
   1165      * shared library whose segments are properly mapped in. */
   1166     TRACE("[ %5d init_library base=0x%08x sz=0x%08x name='%s') ]\n",
   1167           pid, si->base, si->size, si->name);
   1168 
   1169     if(link_image(si, wr_offset)) {
   1170             /* We failed to link.  However, we can only restore libbase
   1171             ** if no additional libraries have moved it since we updated it.
   1172             */
   1173         munmap((void *)si->base, si->size);
   1174         return NULL;
   1175     }
   1176 
   1177     return si;
   1178 }
   1179 
   1180 soinfo *find_library(const char *name)
   1181 {
   1182     soinfo *si;
   1183     const char *bname;
   1184 
   1185 #if ALLOW_SYMBOLS_FROM_MAIN
   1186     if (name == NULL)
   1187         return somain;
   1188 #else
   1189     if (name == NULL)
   1190         return NULL;
   1191 #endif
   1192 
   1193     bname = strrchr(name, '/');
   1194     bname = bname ? bname + 1 : name;
   1195 
   1196     for(si = solist; si != 0; si = si->next){
   1197         if(!strcmp(bname, si->name)) {
   1198             if(si->flags & FLAG_ERROR) {
   1199                 DL_ERR("%5d '%s' failed to load previously", pid, bname);
   1200                 return NULL;
   1201             }
   1202             if(si->flags & FLAG_LINKED) return si;
   1203             DL_ERR("OOPS: %5d recursive link to '%s'", pid, si->name);
   1204             return NULL;
   1205         }
   1206     }
   1207 
   1208     TRACE("[ %5d '%s' has not been loaded yet.  Locating...]\n", pid, name);
   1209     si = load_library(name);
   1210     if(si == NULL)
   1211         return NULL;
   1212     return init_library(si);
   1213 }
   1214 
   1215 /* TODO:
   1216  *   notify gdb of unload
   1217  *   for non-prelinked libraries, find a way to decrement libbase
   1218  */
   1219 static void call_destructors(soinfo *si);
   1220 unsigned unload_library(soinfo *si)
   1221 {
   1222     unsigned *d;
   1223     if (si->refcount == 1) {
   1224         TRACE("%5d unloading '%s'\n", pid, si->name);
   1225         call_destructors(si);
   1226 
   1227         for(d = si->dynamic; *d; d += 2) {
   1228             if(d[0] == DT_NEEDED){
   1229                 soinfo *lsi = (soinfo *)d[1];
   1230                 d[1] = 0;
   1231                 if (validate_soinfo(lsi)) {
   1232                     TRACE("%5d %s needs to unload %s\n", pid,
   1233                           si->name, lsi->name);
   1234                     unload_library(lsi);
   1235                 }
   1236                 else
   1237                     DL_ERR("%5d %s: could not unload dependent library",
   1238                            pid, si->name);
   1239             }
   1240         }
   1241 
   1242         munmap((char *)si->base, si->size);
   1243         notify_gdb_of_unload(si);
   1244         free_info(si);
   1245         si->refcount = 0;
   1246     }
   1247     else {
   1248         si->refcount--;
   1249         PRINT("%5d not unloading '%s', decrementing refcount to %d\n",
   1250               pid, si->name, si->refcount);
   1251     }
   1252     return si->refcount;
   1253 }
   1254 
   1255 /* TODO: don't use unsigned for addrs below. It works, but is not
   1256  * ideal. They should probably be either uint32_t, Elf32_Addr, or unsigned
   1257  * long.
   1258  */
   1259 static int reloc_library(soinfo *si, Elf32_Rel *rel, unsigned count)
   1260 {
   1261     Elf32_Sym *symtab = si->symtab;
   1262     const char *strtab = si->strtab;
   1263     Elf32_Sym *s;
   1264     unsigned base;
   1265     Elf32_Rel *start = rel;
   1266     unsigned idx;
   1267 
   1268     for (idx = 0; idx < count; ++idx) {
   1269         unsigned type = ELF32_R_TYPE(rel->r_info);
   1270         unsigned sym = ELF32_R_SYM(rel->r_info);
   1271         unsigned reloc = (unsigned)(rel->r_offset + si->base);
   1272         unsigned sym_addr = 0;
   1273         char *sym_name = NULL;
   1274 
   1275         DEBUG("%5d Processing '%s' relocation at index %d\n", pid,
   1276               si->name, idx);
   1277         if(sym != 0) {
   1278             sym_name = (char *)(strtab + symtab[sym].st_name);
   1279             s = _do_lookup(si, sym_name, &base);
   1280             if(s == NULL) {
   1281                 /* We only allow an undefined symbol if this is a weak
   1282                    reference..   */
   1283                 s = &symtab[sym];
   1284                 if (ELF32_ST_BIND(s->st_info) != STB_WEAK) {
   1285                     DL_ERR("%5d cannot locate '%s'...\n", pid, sym_name);
   1286                     return -1;
   1287                 }
   1288 
   1289                 /* IHI0044C AAELF 4.5.1.1:
   1290 
   1291                    Libraries are not searched to resolve weak references.
   1292                    It is not an error for a weak reference to remain
   1293                    unsatisfied.
   1294 
   1295                    During linking, the value of an undefined weak reference is:
   1296                    - Zero if the relocation type is absolute
   1297                    - The address of the place if the relocation is pc-relative
   1298                    - The address of nominial base address if the relocation
   1299                      type is base-relative.
   1300                   */
   1301 
   1302                 switch (type) {
   1303 #if defined(ANDROID_ARM_LINKER)
   1304                 case R_ARM_JUMP_SLOT:
   1305                 case R_ARM_GLOB_DAT:
   1306                 case R_ARM_ABS32:
   1307                 case R_ARM_RELATIVE:    /* Don't care. */
   1308                 case R_ARM_NONE:        /* Don't care. */
   1309 #elif defined(ANDROID_X86_LINKER)
   1310                 case R_386_JUMP_SLOT:
   1311                 case R_386_GLOB_DAT:
   1312                 case R_386_32:
   1313                 case R_386_RELATIVE:    /* Dont' care. */
   1314 #endif /* ANDROID_*_LINKER */
   1315                     /* sym_addr was initialized to be zero above or relocation
   1316                        code below does not care about value of sym_addr.
   1317                        No need to do anything.  */
   1318                     break;
   1319 
   1320 #if defined(ANDROID_X86_LINKER)
   1321                 case R_386_PC32:
   1322                     sym_addr = reloc;
   1323                     break;
   1324 #endif /* ANDROID_X86_LINKER */
   1325 
   1326 #if defined(ANDROID_ARM_LINKER)
   1327                 case R_ARM_COPY:
   1328                     /* Fall through.  Can't really copy if weak symbol is
   1329                        not found in run-time.  */
   1330 #endif /* ANDROID_ARM_LINKER */
   1331                 default:
   1332                     DL_ERR("%5d unknown weak reloc type %d @ %p (%d)\n",
   1333                                  pid, type, rel, (int) (rel - start));
   1334                     return -1;
   1335                 }
   1336             } else {
   1337                 /* We got a definition.  */
   1338 #if 0
   1339             if((base == 0) && (si->base != 0)){
   1340                     /* linking from libraries to main image is bad */
   1341                 DL_ERR("%5d cannot locate '%s'...",
   1342                        pid, strtab + symtab[sym].st_name);
   1343                 return -1;
   1344             }
   1345 #endif
   1346                 sym_addr = (unsigned)(s->st_value + base);
   1347 	    }
   1348             COUNT_RELOC(RELOC_SYMBOL);
   1349         } else {
   1350             s = NULL;
   1351         }
   1352 
   1353 /* TODO: This is ugly. Split up the relocations by arch into
   1354  * different files.
   1355  */
   1356         switch(type){
   1357 #if defined(ANDROID_ARM_LINKER)
   1358         case R_ARM_JUMP_SLOT:
   1359             COUNT_RELOC(RELOC_ABSOLUTE);
   1360             MARK(rel->r_offset);
   1361             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1362                        reloc, sym_addr, sym_name);
   1363             *((unsigned*)reloc) = sym_addr;
   1364             break;
   1365         case R_ARM_GLOB_DAT:
   1366             COUNT_RELOC(RELOC_ABSOLUTE);
   1367             MARK(rel->r_offset);
   1368             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1369                        reloc, sym_addr, sym_name);
   1370             *((unsigned*)reloc) = sym_addr;
   1371             break;
   1372         case R_ARM_ABS32:
   1373             COUNT_RELOC(RELOC_ABSOLUTE);
   1374             MARK(rel->r_offset);
   1375             TRACE_TYPE(RELO, "%5d RELO ABS %08x <- %08x %s\n", pid,
   1376                        reloc, sym_addr, sym_name);
   1377             *((unsigned*)reloc) += sym_addr;
   1378             break;
   1379         case R_ARM_REL32:
   1380             COUNT_RELOC(RELOC_RELATIVE);
   1381             MARK(rel->r_offset);
   1382             TRACE_TYPE(RELO, "%5d RELO REL32 %08x <- %08x - %08x %s\n", pid,
   1383                        reloc, sym_addr, rel->r_offset, sym_name);
   1384             *((unsigned*)reloc) += sym_addr - rel->r_offset;
   1385             break;
   1386 #elif defined(ANDROID_X86_LINKER)
   1387         case R_386_JUMP_SLOT:
   1388             COUNT_RELOC(RELOC_ABSOLUTE);
   1389             MARK(rel->r_offset);
   1390             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1391                        reloc, sym_addr, sym_name);
   1392             *((unsigned*)reloc) = sym_addr;
   1393             break;
   1394         case R_386_GLOB_DAT:
   1395             COUNT_RELOC(RELOC_ABSOLUTE);
   1396             MARK(rel->r_offset);
   1397             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1398                        reloc, sym_addr, sym_name);
   1399             *((unsigned*)reloc) = sym_addr;
   1400             break;
   1401 #endif /* ANDROID_*_LINKER */
   1402 
   1403 #if defined(ANDROID_ARM_LINKER)
   1404         case R_ARM_RELATIVE:
   1405 #elif defined(ANDROID_X86_LINKER)
   1406         case R_386_RELATIVE:
   1407 #endif /* ANDROID_*_LINKER */
   1408             COUNT_RELOC(RELOC_RELATIVE);
   1409             MARK(rel->r_offset);
   1410             if(sym){
   1411                 DL_ERR("%5d odd RELATIVE form...", pid);
   1412                 return -1;
   1413             }
   1414             TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", pid,
   1415                        reloc, si->base);
   1416             *((unsigned*)reloc) += si->base;
   1417             break;
   1418 
   1419 #if defined(ANDROID_X86_LINKER)
   1420         case R_386_32:
   1421             COUNT_RELOC(RELOC_RELATIVE);
   1422             MARK(rel->r_offset);
   1423 
   1424             TRACE_TYPE(RELO, "%5d RELO R_386_32 %08x <- +%08x %s\n", pid,
   1425                        reloc, sym_addr, sym_name);
   1426             *((unsigned *)reloc) += (unsigned)sym_addr;
   1427             break;
   1428 
   1429         case R_386_PC32:
   1430             COUNT_RELOC(RELOC_RELATIVE);
   1431             MARK(rel->r_offset);
   1432             TRACE_TYPE(RELO, "%5d RELO R_386_PC32 %08x <- "
   1433                        "+%08x (%08x - %08x) %s\n", pid, reloc,
   1434                        (sym_addr - reloc), sym_addr, reloc, sym_name);
   1435             *((unsigned *)reloc) += (unsigned)(sym_addr - reloc);
   1436             break;
   1437 #endif /* ANDROID_X86_LINKER */
   1438 
   1439 #ifdef ANDROID_ARM_LINKER
   1440         case R_ARM_COPY:
   1441             COUNT_RELOC(RELOC_COPY);
   1442             MARK(rel->r_offset);
   1443             TRACE_TYPE(RELO, "%5d RELO %08x <- %d @ %08x %s\n", pid,
   1444                        reloc, s->st_size, sym_addr, sym_name);
   1445             memcpy((void*)reloc, (void*)sym_addr, s->st_size);
   1446             break;
   1447         case R_ARM_NONE:
   1448             break;
   1449 #endif /* ANDROID_ARM_LINKER */
   1450 
   1451         default:
   1452             DL_ERR("%5d unknown reloc type %d @ %p (%d)",
   1453                   pid, type, rel, (int) (rel - start));
   1454             return -1;
   1455         }
   1456         rel++;
   1457     }
   1458     return 0;
   1459 }
   1460 
   1461 #if defined(ANDROID_SH_LINKER)
   1462 static int reloc_library_a(soinfo *si, Elf32_Rela *rela, unsigned count)
   1463 {
   1464     Elf32_Sym *symtab = si->symtab;
   1465     const char *strtab = si->strtab;
   1466     Elf32_Sym *s;
   1467     unsigned base;
   1468     Elf32_Rela *start = rela;
   1469     unsigned idx;
   1470 
   1471     for (idx = 0; idx < count; ++idx) {
   1472         unsigned type = ELF32_R_TYPE(rela->r_info);
   1473         unsigned sym = ELF32_R_SYM(rela->r_info);
   1474         unsigned reloc = (unsigned)(rela->r_offset + si->base);
   1475         unsigned sym_addr = 0;
   1476         char *sym_name = NULL;
   1477 
   1478         DEBUG("%5d Processing '%s' relocation at index %d\n", pid,
   1479               si->name, idx);
   1480         if(sym != 0) {
   1481             sym_name = (char *)(strtab + symtab[sym].st_name);
   1482             s = _do_lookup(si, sym_name, &base);
   1483             if(s == 0) {
   1484                 DL_ERR("%5d cannot locate '%s'...", pid, sym_name);
   1485                 return -1;
   1486             }
   1487 #if 0
   1488             if((base == 0) && (si->base != 0)){
   1489                     /* linking from libraries to main image is bad */
   1490                 DL_ERR("%5d cannot locate '%s'...",
   1491                        pid, strtab + symtab[sym].st_name);
   1492                 return -1;
   1493             }
   1494 #endif
   1495             if ((s->st_shndx == SHN_UNDEF) && (s->st_value != 0)) {
   1496                 DL_ERR("%5d In '%s', shndx=%d && value=0x%08x. We do not "
   1497                       "handle this yet", pid, si->name, s->st_shndx,
   1498                       s->st_value);
   1499                 return -1;
   1500             }
   1501             sym_addr = (unsigned)(s->st_value + base);
   1502             COUNT_RELOC(RELOC_SYMBOL);
   1503         } else {
   1504             s = 0;
   1505         }
   1506 
   1507 /* TODO: This is ugly. Split up the relocations by arch into
   1508  * different files.
   1509  */
   1510         switch(type){
   1511         case R_SH_JUMP_SLOT:
   1512             COUNT_RELOC(RELOC_ABSOLUTE);
   1513             MARK(rela->r_offset);
   1514             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1515                        reloc, sym_addr, sym_name);
   1516             *((unsigned*)reloc) = sym_addr;
   1517             break;
   1518         case R_SH_GLOB_DAT:
   1519             COUNT_RELOC(RELOC_ABSOLUTE);
   1520             MARK(rela->r_offset);
   1521             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1522                        reloc, sym_addr, sym_name);
   1523             *((unsigned*)reloc) = sym_addr;
   1524             break;
   1525         case R_SH_DIR32:
   1526             COUNT_RELOC(RELOC_ABSOLUTE);
   1527             MARK(rela->r_offset);
   1528             TRACE_TYPE(RELO, "%5d RELO DIR32 %08x <- %08x %s\n", pid,
   1529                        reloc, sym_addr, sym_name);
   1530             *((unsigned*)reloc) += sym_addr;
   1531             break;
   1532         case R_SH_RELATIVE:
   1533             COUNT_RELOC(RELOC_RELATIVE);
   1534             MARK(rela->r_offset);
   1535             if(sym){
   1536                 DL_ERR("%5d odd RELATIVE form...", pid);
   1537                 return -1;
   1538             }
   1539             TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", pid,
   1540                        reloc, si->base);
   1541             *((unsigned*)reloc) += si->base;
   1542             break;
   1543 
   1544         default:
   1545             DL_ERR("%5d unknown reloc type %d @ %p (%d)",
   1546                   pid, type, rela, (int) (rela - start));
   1547             return -1;
   1548         }
   1549         rela++;
   1550     }
   1551     return 0;
   1552 }
   1553 #endif /* ANDROID_SH_LINKER */
   1554 
   1555 
   1556 /* Please read the "Initialization and Termination functions" functions.
   1557  * of the linker design note in bionic/linker/README.TXT to understand
   1558  * what the following code is doing.
   1559  *
   1560  * The important things to remember are:
   1561  *
   1562  *   DT_PREINIT_ARRAY must be called first for executables, and should
   1563  *   not appear in shared libraries.
   1564  *
   1565  *   DT_INIT should be called before DT_INIT_ARRAY if both are present
   1566  *
   1567  *   DT_FINI should be called after DT_FINI_ARRAY if both are present
   1568  *
   1569  *   DT_FINI_ARRAY must be parsed in reverse order.
   1570  */
   1571 
   1572 static void call_array(unsigned *ctor, int count, int reverse)
   1573 {
   1574     int n, inc = 1;
   1575 
   1576     if (reverse) {
   1577         ctor += (count-1);
   1578         inc   = -1;
   1579     }
   1580 
   1581     for(n = count; n > 0; n--) {
   1582         TRACE("[ %5d Looking at %s *0x%08x == 0x%08x ]\n", pid,
   1583               reverse ? "dtor" : "ctor",
   1584               (unsigned)ctor, (unsigned)*ctor);
   1585         void (*func)() = (void (*)()) *ctor;
   1586         ctor += inc;
   1587         if(((int) func == 0) || ((int) func == -1)) continue;
   1588         TRACE("[ %5d Calling func @ 0x%08x ]\n", pid, (unsigned)func);
   1589         func();
   1590     }
   1591 }
   1592 
   1593 static void call_constructors(soinfo *si)
   1594 {
   1595     if (si->flags & FLAG_EXE) {
   1596         TRACE("[ %5d Calling preinit_array @ 0x%08x [%d] for '%s' ]\n",
   1597               pid, (unsigned)si->preinit_array, si->preinit_array_count,
   1598               si->name);
   1599         call_array(si->preinit_array, si->preinit_array_count, 0);
   1600         TRACE("[ %5d Done calling preinit_array for '%s' ]\n", pid, si->name);
   1601     } else {
   1602         if (si->preinit_array) {
   1603             DL_ERR("%5d Shared library '%s' has a preinit_array table @ 0x%08x."
   1604                    " This is INVALID.", pid, si->name,
   1605                    (unsigned)si->preinit_array);
   1606         }
   1607     }
   1608 
   1609     if (si->init_func) {
   1610         TRACE("[ %5d Calling init_func @ 0x%08x for '%s' ]\n", pid,
   1611               (unsigned)si->init_func, si->name);
   1612         si->init_func();
   1613         TRACE("[ %5d Done calling init_func for '%s' ]\n", pid, si->name);
   1614     }
   1615 
   1616     if (si->init_array) {
   1617         TRACE("[ %5d Calling init_array @ 0x%08x [%d] for '%s' ]\n", pid,
   1618               (unsigned)si->init_array, si->init_array_count, si->name);
   1619         call_array(si->init_array, si->init_array_count, 0);
   1620         TRACE("[ %5d Done calling init_array for '%s' ]\n", pid, si->name);
   1621     }
   1622 }
   1623 
   1624 
   1625 static void call_destructors(soinfo *si)
   1626 {
   1627     if (si->fini_array) {
   1628         TRACE("[ %5d Calling fini_array @ 0x%08x [%d] for '%s' ]\n", pid,
   1629               (unsigned)si->fini_array, si->fini_array_count, si->name);
   1630         call_array(si->fini_array, si->fini_array_count, 1);
   1631         TRACE("[ %5d Done calling fini_array for '%s' ]\n", pid, si->name);
   1632     }
   1633 
   1634     if (si->fini_func) {
   1635         TRACE("[ %5d Calling fini_func @ 0x%08x for '%s' ]\n", pid,
   1636               (unsigned)si->fini_func, si->name);
   1637         si->fini_func();
   1638         TRACE("[ %5d Done calling fini_func for '%s' ]\n", pid, si->name);
   1639     }
   1640 }
   1641 
   1642 /* Force any of the closed stdin, stdout and stderr to be associated with
   1643    /dev/null. */
   1644 static int nullify_closed_stdio (void)
   1645 {
   1646     int dev_null, i, status;
   1647     int return_value = 0;
   1648 
   1649     dev_null = open("/dev/null", O_RDWR);
   1650     if (dev_null < 0) {
   1651         DL_ERR("Cannot open /dev/null.");
   1652         return -1;
   1653     }
   1654     TRACE("[ %5d Opened /dev/null file-descriptor=%d]\n", pid, dev_null);
   1655 
   1656     /* If any of the stdio file descriptors is valid and not associated
   1657        with /dev/null, dup /dev/null to it.  */
   1658     for (i = 0; i < 3; i++) {
   1659         /* If it is /dev/null already, we are done. */
   1660         if (i == dev_null)
   1661             continue;
   1662 
   1663         TRACE("[ %5d Nullifying stdio file descriptor %d]\n", pid, i);
   1664         /* The man page of fcntl does not say that fcntl(..,F_GETFL)
   1665            can be interrupted but we do this just to be safe. */
   1666         do {
   1667           status = fcntl(i, F_GETFL);
   1668         } while (status < 0 && errno == EINTR);
   1669 
   1670         /* If file is openned, we are good. */
   1671         if (status >= 0)
   1672           continue;
   1673 
   1674         /* The only error we allow is that the file descriptor does not
   1675            exist, in which case we dup /dev/null to it. */
   1676         if (errno != EBADF) {
   1677             DL_ERR("nullify_stdio: unhandled error %s", strerror(errno));
   1678             return_value = -1;
   1679             continue;
   1680         }
   1681 
   1682         /* Try dupping /dev/null to this stdio file descriptor and
   1683            repeat if there is a signal.  Note that any errors in closing
   1684            the stdio descriptor are lost.  */
   1685         do {
   1686             status = dup2(dev_null, i);
   1687         } while (status < 0 && errno == EINTR);
   1688 
   1689         if (status < 0) {
   1690             DL_ERR("nullify_stdio: dup2 error %s", strerror(errno));
   1691             return_value = -1;
   1692             continue;
   1693         }
   1694     }
   1695 
   1696     /* If /dev/null is not one of the stdio file descriptors, close it. */
   1697     if (dev_null > 2) {
   1698         TRACE("[ %5d Closing /dev/null file-descriptor=%d]\n", pid, dev_null);
   1699         do {
   1700             status = close(dev_null);
   1701         } while (status < 0 && errno == EINTR);
   1702 
   1703         if (status < 0) {
   1704             DL_ERR("nullify_stdio: close error %s", strerror(errno));
   1705             return_value = -1;
   1706         }
   1707     }
   1708 
   1709     return return_value;
   1710 }
   1711 
   1712 static int link_image(soinfo *si, unsigned wr_offset)
   1713 {
   1714     unsigned *d;
   1715     Elf32_Phdr *phdr = si->phdr;
   1716     int phnum = si->phnum;
   1717 
   1718     INFO("[ %5d linking %s ]\n", pid, si->name);
   1719     DEBUG("%5d si->base = 0x%08x si->flags = 0x%08x\n", pid,
   1720           si->base, si->flags);
   1721 
   1722     if (si->flags & FLAG_EXE) {
   1723         /* Locate the needed program segments (DYNAMIC/ARM_EXIDX) for
   1724          * linkage info if this is the executable. If this was a
   1725          * dynamic lib, that would have been done at load time.
   1726          *
   1727          * TODO: It's unfortunate that small pieces of this are
   1728          * repeated from the load_library routine. Refactor this just
   1729          * slightly to reuse these bits.
   1730          */
   1731         si->size = 0;
   1732         for(; phnum > 0; --phnum, ++phdr) {
   1733 #ifdef ANDROID_ARM_LINKER
   1734             if(phdr->p_type == PT_ARM_EXIDX) {
   1735                 /* exidx entries (used for stack unwinding) are 8 bytes each.
   1736                  */
   1737                 si->ARM_exidx = (unsigned *)phdr->p_vaddr;
   1738                 si->ARM_exidx_count = phdr->p_memsz / 8;
   1739             }
   1740 #endif
   1741             if (phdr->p_type == PT_LOAD) {
   1742                 /* For the executable, we use the si->size field only in
   1743                    dl_unwind_find_exidx(), so the meaning of si->size
   1744                    is not the size of the executable; it is the last
   1745                    virtual address of the loadable part of the executable;
   1746                    since si->base == 0 for an executable, we use the
   1747                    range [0, si->size) to determine whether a PC value
   1748                    falls within the executable section.  Of course, if
   1749                    a value is below phdr->p_vaddr, it's not in the
   1750                    executable section, but a) we shouldn't be asking for
   1751                    such a value anyway, and b) if we have to provide
   1752                    an EXIDX for such a value, then the executable's
   1753                    EXIDX is probably the better choice.
   1754                 */
   1755                 DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
   1756                 if (phdr->p_vaddr + phdr->p_memsz > si->size)
   1757                     si->size = phdr->p_vaddr + phdr->p_memsz;
   1758                 /* try to remember what range of addresses should be write
   1759                  * protected */
   1760                 if (!(phdr->p_flags & PF_W)) {
   1761                     unsigned _end;
   1762 
   1763                     if (phdr->p_vaddr < si->wrprotect_start)
   1764                         si->wrprotect_start = phdr->p_vaddr;
   1765                     _end = (((phdr->p_vaddr + phdr->p_memsz + PAGE_SIZE - 1) &
   1766                              (~PAGE_MASK)));
   1767                     if (_end > si->wrprotect_end)
   1768                         si->wrprotect_end = _end;
   1769                 }
   1770             } else if (phdr->p_type == PT_DYNAMIC) {
   1771                 if (si->dynamic != (unsigned *)-1) {
   1772                     DL_ERR("%5d multiple PT_DYNAMIC segments found in '%s'. "
   1773                           "Segment at 0x%08x, previously one found at 0x%08x",
   1774                           pid, si->name, si->base + phdr->p_vaddr,
   1775                           (unsigned)si->dynamic);
   1776                     goto fail;
   1777                 }
   1778                 DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
   1779                 si->dynamic = (unsigned *) (si->base + phdr->p_vaddr);
   1780             }
   1781         }
   1782     }
   1783 
   1784     if (si->dynamic == (unsigned *)-1) {
   1785         DL_ERR("%5d missing PT_DYNAMIC?!", pid);
   1786         goto fail;
   1787     }
   1788 
   1789     DEBUG("%5d dynamic = %p\n", pid, si->dynamic);
   1790 
   1791     /* extract useful information from dynamic section */
   1792     for(d = si->dynamic; *d; d++){
   1793         DEBUG("%5d d = %p, d[0] = 0x%08x d[1] = 0x%08x\n", pid, d, d[0], d[1]);
   1794         switch(*d++){
   1795         case DT_HASH:
   1796             si->nbucket = ((unsigned *) (si->base + *d))[0];
   1797             si->nchain = ((unsigned *) (si->base + *d))[1];
   1798             si->bucket = (unsigned *) (si->base + *d + 8);
   1799             si->chain = (unsigned *) (si->base + *d + 8 + si->nbucket * 4);
   1800             break;
   1801         case DT_STRTAB:
   1802             si->strtab = (const char *) (si->base + *d);
   1803             break;
   1804         case DT_SYMTAB:
   1805             si->symtab = (Elf32_Sym *) (si->base + *d);
   1806             break;
   1807 #if !defined(ANDROID_SH_LINKER)
   1808         case DT_PLTREL:
   1809             if(*d != DT_REL) {
   1810                 DL_ERR("DT_RELA not supported");
   1811                 goto fail;
   1812             }
   1813             break;
   1814 #endif
   1815 #ifdef ANDROID_SH_LINKER
   1816         case DT_JMPREL:
   1817             si->plt_rela = (Elf32_Rela*) (si->base + *d);
   1818             break;
   1819         case DT_PLTRELSZ:
   1820             si->plt_rela_count = *d / sizeof(Elf32_Rela);
   1821             break;
   1822 #else
   1823         case DT_JMPREL:
   1824             si->plt_rel = (Elf32_Rel*) (si->base + *d);
   1825             break;
   1826         case DT_PLTRELSZ:
   1827             si->plt_rel_count = *d / 8;
   1828             break;
   1829 #endif
   1830         case DT_REL:
   1831             si->rel = (Elf32_Rel*) (si->base + *d);
   1832             break;
   1833         case DT_RELSZ:
   1834             si->rel_count = *d / 8;
   1835             break;
   1836 #ifdef ANDROID_SH_LINKER
   1837         case DT_RELASZ:
   1838             si->rela_count = *d / sizeof(Elf32_Rela);
   1839              break;
   1840 #endif
   1841         case DT_PLTGOT:
   1842             /* Save this in case we decide to do lazy binding. We don't yet. */
   1843             si->plt_got = (unsigned *)(si->base + *d);
   1844             break;
   1845         case DT_DEBUG:
   1846             // Set the DT_DEBUG entry to the addres of _r_debug for GDB
   1847             *d = (int) &_r_debug;
   1848             break;
   1849 #ifdef ANDROID_SH_LINKER
   1850         case DT_RELA:
   1851             si->rela = (Elf32_Rela *) (si->base + *d);
   1852             break;
   1853 #else
   1854          case DT_RELA:
   1855             DL_ERR("%5d DT_RELA not supported", pid);
   1856             goto fail;
   1857 #endif
   1858         case DT_INIT:
   1859             si->init_func = (void (*)(void))(si->base + *d);
   1860             DEBUG("%5d %s constructors (init func) found at %p\n",
   1861                   pid, si->name, si->init_func);
   1862             break;
   1863         case DT_FINI:
   1864             si->fini_func = (void (*)(void))(si->base + *d);
   1865             DEBUG("%5d %s destructors (fini func) found at %p\n",
   1866                   pid, si->name, si->fini_func);
   1867             break;
   1868         case DT_INIT_ARRAY:
   1869             si->init_array = (unsigned *)(si->base + *d);
   1870             DEBUG("%5d %s constructors (init_array) found at %p\n",
   1871                   pid, si->name, si->init_array);
   1872             break;
   1873         case DT_INIT_ARRAYSZ:
   1874             si->init_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1875             break;
   1876         case DT_FINI_ARRAY:
   1877             si->fini_array = (unsigned *)(si->base + *d);
   1878             DEBUG("%5d %s destructors (fini_array) found at %p\n",
   1879                   pid, si->name, si->fini_array);
   1880             break;
   1881         case DT_FINI_ARRAYSZ:
   1882             si->fini_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1883             break;
   1884         case DT_PREINIT_ARRAY:
   1885             si->preinit_array = (unsigned *)(si->base + *d);
   1886             DEBUG("%5d %s constructors (preinit_array) found at %p\n",
   1887                   pid, si->name, si->preinit_array);
   1888             break;
   1889         case DT_PREINIT_ARRAYSZ:
   1890             si->preinit_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1891             break;
   1892         case DT_TEXTREL:
   1893             /* TODO: make use of this. */
   1894             /* this means that we might have to write into where the text
   1895              * segment was loaded during relocation... Do something with
   1896              * it.
   1897              */
   1898             DEBUG("%5d Text segment should be writable during relocation.\n",
   1899                   pid);
   1900             break;
   1901         }
   1902     }
   1903 
   1904     DEBUG("%5d si->base = 0x%08x, si->strtab = %p, si->symtab = %p\n",
   1905            pid, si->base, si->strtab, si->symtab);
   1906 
   1907     if((si->strtab == 0) || (si->symtab == 0)) {
   1908         DL_ERR("%5d missing essential tables", pid);
   1909         goto fail;
   1910     }
   1911 
   1912     /* if this is the main executable, then load all of the preloads now */
   1913     if(si->flags & FLAG_EXE) {
   1914         int i;
   1915         memset(preloads, 0, sizeof(preloads));
   1916         for(i = 0; ldpreload_names[i] != NULL; i++) {
   1917             soinfo *lsi = find_library(ldpreload_names[i]);
   1918             if(lsi == 0) {
   1919                 strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
   1920                 DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
   1921                        pid, ldpreload_names[i], si->name, tmp_err_buf);
   1922                 goto fail;
   1923             }
   1924             lsi->refcount++;
   1925             preloads[i] = lsi;
   1926         }
   1927     }
   1928 
   1929     for(d = si->dynamic; *d; d += 2) {
   1930         if(d[0] == DT_NEEDED){
   1931             DEBUG("%5d %s needs %s\n", pid, si->name, si->strtab + d[1]);
   1932             soinfo *lsi = find_library(si->strtab + d[1]);
   1933             if(lsi == 0) {
   1934                 strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
   1935                 DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
   1936                        pid, si->strtab + d[1], si->name, tmp_err_buf);
   1937                 goto fail;
   1938             }
   1939             /* Save the soinfo of the loaded DT_NEEDED library in the payload
   1940                of the DT_NEEDED entry itself, so that we can retrieve the
   1941                soinfo directly later from the dynamic segment.  This is a hack,
   1942                but it allows us to map from DT_NEEDED to soinfo efficiently
   1943                later on when we resolve relocations, trying to look up a symgol
   1944                with dlsym().
   1945             */
   1946             d[1] = (unsigned)lsi;
   1947             lsi->refcount++;
   1948         }
   1949     }
   1950 
   1951     if(si->plt_rel) {
   1952         DEBUG("[ %5d relocating %s plt ]\n", pid, si->name );
   1953         if(reloc_library(si, si->plt_rel, si->plt_rel_count))
   1954             goto fail;
   1955     }
   1956     if(si->rel) {
   1957         DEBUG("[ %5d relocating %s ]\n", pid, si->name );
   1958         if(reloc_library(si, si->rel, si->rel_count))
   1959             goto fail;
   1960     }
   1961 
   1962 #ifdef ANDROID_SH_LINKER
   1963     if(si->plt_rela) {
   1964         DEBUG("[ %5d relocating %s plt ]\n", pid, si->name );
   1965         if(reloc_library_a(si, si->plt_rela, si->plt_rela_count))
   1966             goto fail;
   1967     }
   1968     if(si->rela) {
   1969         DEBUG("[ %5d relocating %s ]\n", pid, si->name );
   1970         if(reloc_library_a(si, si->rela, si->rela_count))
   1971             goto fail;
   1972     }
   1973 #endif /* ANDROID_SH_LINKER */
   1974 
   1975     si->flags |= FLAG_LINKED;
   1976     DEBUG("[ %5d finished linking %s ]\n", pid, si->name);
   1977 
   1978 #if 0
   1979     /* This is the way that the old dynamic linker did protection of
   1980      * non-writable areas. It would scan section headers and find where
   1981      * .text ended (rather where .data/.bss began) and assume that this is
   1982      * the upper range of the non-writable area. This is too coarse,
   1983      * and is kept here for reference until we fully move away from single
   1984      * segment elf objects. See the code in get_wr_offset (also #if'd 0)
   1985      * that made this possible.
   1986      */
   1987     if(wr_offset < 0xffffffff){
   1988         mprotect((void*) si->base, wr_offset, PROT_READ | PROT_EXEC);
   1989     }
   1990 #else
   1991     /* TODO: Verify that this does the right thing in all cases, as it
   1992      * presently probably does not. It is possible that an ELF image will
   1993      * come with multiple read-only segments. What we ought to do is scan
   1994      * the program headers again and mprotect all the read-only segments.
   1995      * To prevent re-scanning the program header, we would have to build a
   1996      * list of loadable segments in si, and then scan that instead. */
   1997     if (si->wrprotect_start != 0xffffffff && si->wrprotect_end != 0) {
   1998         mprotect((void *)si->wrprotect_start,
   1999                  si->wrprotect_end - si->wrprotect_start,
   2000                  PROT_READ | PROT_EXEC);
   2001     }
   2002 #endif
   2003 
   2004     /* If this is a SET?ID program, dup /dev/null to opened stdin,
   2005        stdout and stderr to close a security hole described in:
   2006 
   2007     ftp://ftp.freebsd.org/pub/FreeBSD/CERT/advisories/FreeBSD-SA-02:23.stdio.asc
   2008 
   2009      */
   2010     if (program_is_setuid)
   2011         nullify_closed_stdio ();
   2012     notify_gdb_of_load(si);
   2013     call_constructors(si);
   2014     return 0;
   2015 
   2016 fail:
   2017     ERROR("failed to link %s\n", si->name);
   2018     si->flags |= FLAG_ERROR;
   2019     return -1;
   2020 }
   2021 
   2022 static void parse_library_path(const char *path, char *delim)
   2023 {
   2024     size_t len;
   2025     char *ldpaths_bufp = ldpaths_buf;
   2026     int i = 0;
   2027 
   2028     len = strlcpy(ldpaths_buf, path, sizeof(ldpaths_buf));
   2029 
   2030     while (i < LDPATH_MAX && (ldpaths[i] = strsep(&ldpaths_bufp, delim))) {
   2031         if (*ldpaths[i] != '\0')
   2032             ++i;
   2033     }
   2034 
   2035     /* Forget the last path if we had to truncate; this occurs if the 2nd to
   2036      * last char isn't '\0' (i.e. not originally a delim). */
   2037     if (i > 0 && len >= sizeof(ldpaths_buf) &&
   2038             ldpaths_buf[sizeof(ldpaths_buf) - 2] != '\0') {
   2039         ldpaths[i - 1] = NULL;
   2040     } else {
   2041         ldpaths[i] = NULL;
   2042     }
   2043 }
   2044 
   2045 static void parse_preloads(const char *path, char *delim)
   2046 {
   2047     size_t len;
   2048     char *ldpreloads_bufp = ldpreloads_buf;
   2049     int i = 0;
   2050 
   2051     len = strlcpy(ldpreloads_buf, path, sizeof(ldpreloads_buf));
   2052 
   2053     while (i < LDPRELOAD_MAX && (ldpreload_names[i] = strsep(&ldpreloads_bufp, delim))) {
   2054         if (*ldpreload_names[i] != '\0') {
   2055             ++i;
   2056         }
   2057     }
   2058 
   2059     /* Forget the last path if we had to truncate; this occurs if the 2nd to
   2060      * last char isn't '\0' (i.e. not originally a delim). */
   2061     if (i > 0 && len >= sizeof(ldpreloads_buf) &&
   2062             ldpreloads_buf[sizeof(ldpreloads_buf) - 2] != '\0') {
   2063         ldpreload_names[i - 1] = NULL;
   2064     } else {
   2065         ldpreload_names[i] = NULL;
   2066     }
   2067 }
   2068 
   2069 int main(int argc, char **argv)
   2070 {
   2071     return 0;
   2072 }
   2073 
   2074 #define ANDROID_TLS_SLOTS  BIONIC_TLS_SLOTS
   2075 
   2076 static void * __tls_area[ANDROID_TLS_SLOTS];
   2077 
   2078 unsigned __linker_init(unsigned **elfdata)
   2079 {
   2080     static soinfo linker_soinfo;
   2081 
   2082     int argc = (int) *elfdata;
   2083     char **argv = (char**) (elfdata + 1);
   2084     unsigned *vecs = (unsigned*) (argv + argc + 1);
   2085     soinfo *si;
   2086     struct link_map * map;
   2087     const char *ldpath_env = NULL;
   2088     const char *ldpreload_env = NULL;
   2089 
   2090     /* Setup a temporary TLS area that is used to get a working
   2091      * errno for system calls.
   2092      */
   2093     __set_tls(__tls_area);
   2094 
   2095     pid = getpid();
   2096 
   2097 #if TIMING
   2098     struct timeval t0, t1;
   2099     gettimeofday(&t0, 0);
   2100 #endif
   2101 
   2102     /* NOTE: we store the elfdata pointer on a special location
   2103      *       of the temporary TLS area in order to pass it to
   2104      *       the C Library's runtime initializer.
   2105      *
   2106      *       The initializer must clear the slot and reset the TLS
   2107      *       to point to a different location to ensure that no other
   2108      *       shared library constructor can access it.
   2109      */
   2110     __tls_area[TLS_SLOT_BIONIC_PREINIT] = elfdata;
   2111 
   2112     /* Are we setuid? */
   2113     program_is_setuid = (getuid() != geteuid()) || (getgid() != getegid());
   2114 
   2115     /* Initialize environment functions, and get to the ELF aux vectors table */
   2116     vecs = linker_env_init(vecs);
   2117 
   2118     /* Sanitize environment if we're loading a setuid program */
   2119     if (program_is_setuid)
   2120         linker_env_secure();
   2121 
   2122     debugger_init();
   2123 
   2124     /* Get a few environment variables */
   2125     {
   2126         const char* env;
   2127         env = linker_env_get("DEBUG"); /* XXX: TODO: Change to LD_DEBUG */
   2128         if (env)
   2129             debug_verbosity = atoi(env);
   2130 
   2131         /* Normally, these are cleaned by linker_env_secure, but the test
   2132          * against program_is_setuid doesn't cost us anything */
   2133         if (!program_is_setuid) {
   2134             ldpath_env = linker_env_get("LD_LIBRARY_PATH");
   2135             ldpreload_env = linker_env_get("LD_PRELOAD");
   2136         }
   2137     }
   2138 
   2139     INFO("[ android linker & debugger ]\n");
   2140     DEBUG("%5d elfdata @ 0x%08x\n", pid, (unsigned)elfdata);
   2141 
   2142     si = alloc_info(argv[0]);
   2143     if(si == 0) {
   2144         exit(-1);
   2145     }
   2146 
   2147         /* bootstrap the link map, the main exe always needs to be first */
   2148     si->flags |= FLAG_EXE;
   2149     map = &(si->linkmap);
   2150 
   2151     map->l_addr = 0;
   2152     map->l_name = argv[0];
   2153     map->l_prev = NULL;
   2154     map->l_next = NULL;
   2155 
   2156     _r_debug.r_map = map;
   2157     r_debug_tail = map;
   2158 
   2159         /* gdb expects the linker to be in the debug shared object list,
   2160          * and we need to make sure that the reported load address is zero.
   2161          * Without this, gdb gets the wrong idea of where rtld_db_dlactivity()
   2162          * is.  Don't use alloc_info(), because the linker shouldn't
   2163          * be on the soinfo list.
   2164          */
   2165     strlcpy((char*) linker_soinfo.name, "/system/bin/linker", sizeof linker_soinfo.name);
   2166     linker_soinfo.flags = 0;
   2167     linker_soinfo.base = 0;     // This is the important part; must be zero.
   2168     insert_soinfo_into_debug_map(&linker_soinfo);
   2169 
   2170         /* extract information passed from the kernel */
   2171     while(vecs[0] != 0){
   2172         switch(vecs[0]){
   2173         case AT_PHDR:
   2174             si->phdr = (Elf32_Phdr*) vecs[1];
   2175             break;
   2176         case AT_PHNUM:
   2177             si->phnum = (int) vecs[1];
   2178             break;
   2179         case AT_ENTRY:
   2180             si->entry = vecs[1];
   2181             break;
   2182         }
   2183         vecs += 2;
   2184     }
   2185 
   2186     si->base = 0;
   2187     si->dynamic = (unsigned *)-1;
   2188     si->wrprotect_start = 0xffffffff;
   2189     si->wrprotect_end = 0;
   2190     si->refcount = 1;
   2191 
   2192         /* Use LD_LIBRARY_PATH if we aren't setuid/setgid */
   2193     if (ldpath_env)
   2194         parse_library_path(ldpath_env, ":");
   2195 
   2196     if (ldpreload_env) {
   2197         parse_preloads(ldpreload_env, " :");
   2198     }
   2199 
   2200     if(link_image(si, 0)) {
   2201         char errmsg[] = "CANNOT LINK EXECUTABLE\n";
   2202         write(2, __linker_dl_err_buf, strlen(__linker_dl_err_buf));
   2203         write(2, errmsg, sizeof(errmsg));
   2204         exit(-1);
   2205     }
   2206 
   2207 #if ALLOW_SYMBOLS_FROM_MAIN
   2208     /* Set somain after we've loaded all the libraries in order to prevent
   2209      * linking of symbols back to the main image, which is not set up at that
   2210      * point yet.
   2211      */
   2212     somain = si;
   2213 #endif
   2214 
   2215 #if TIMING
   2216     gettimeofday(&t1,NULL);
   2217     PRINT("LINKER TIME: %s: %d microseconds\n", argv[0], (int) (
   2218                (((long long)t1.tv_sec * 1000000LL) + (long long)t1.tv_usec) -
   2219                (((long long)t0.tv_sec * 1000000LL) + (long long)t0.tv_usec)
   2220                ));
   2221 #endif
   2222 #if STATS
   2223     PRINT("RELO STATS: %s: %d abs, %d rel, %d copy, %d symbol\n", argv[0],
   2224            linker_stats.reloc[RELOC_ABSOLUTE],
   2225            linker_stats.reloc[RELOC_RELATIVE],
   2226            linker_stats.reloc[RELOC_COPY],
   2227            linker_stats.reloc[RELOC_SYMBOL]);
   2228 #endif
   2229 #if COUNT_PAGES
   2230     {
   2231         unsigned n;
   2232         unsigned i;
   2233         unsigned count = 0;
   2234         for(n = 0; n < 4096; n++){
   2235             if(bitmask[n]){
   2236                 unsigned x = bitmask[n];
   2237                 for(i = 0; i < 8; i++){
   2238                     if(x & 1) count++;
   2239                     x >>= 1;
   2240                 }
   2241             }
   2242         }
   2243         PRINT("PAGES MODIFIED: %s: %d (%dKB)\n", argv[0], count, count * 4);
   2244     }
   2245 #endif
   2246 
   2247 #if TIMING || STATS || COUNT_PAGES
   2248     fflush(stdout);
   2249 #endif
   2250 
   2251     TRACE("[ %5d Ready to execute '%s' @ 0x%08x ]\n", pid, si->name,
   2252           si->entry);
   2253     return si->entry;
   2254 }
   2255