Home | History | Annotate | Download | only in linker
      1 /*
      2  * Copyright (C) 2008, 2009 The Android Open Source Project
      3  * All rights reserved.
      4  *
      5  * Redistribution and use in source and binary forms, with or without
      6  * modification, are permitted provided that the following conditions
      7  * are met:
      8  *  * Redistributions of source code must retain the above copyright
      9  *    notice, this list of conditions and the following disclaimer.
     10  *  * Redistributions in binary form must reproduce the above copyright
     11  *    notice, this list of conditions and the following disclaimer in
     12  *    the documentation and/or other materials provided with the
     13  *    distribution.
     14  *
     15  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     16  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     17  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
     18  * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
     19  * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
     20  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
     21  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
     22  * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
     23  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
     24  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
     25  * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     26  * SUCH DAMAGE.
     27  */
     28 
     29 #include <linux/auxvec.h>
     30 
     31 #include <stdio.h>
     32 #include <stdlib.h>
     33 #include <string.h>
     34 #include <unistd.h>
     35 #include <fcntl.h>
     36 #include <errno.h>
     37 #include <dlfcn.h>
     38 #include <sys/stat.h>
     39 
     40 #include <pthread.h>
     41 
     42 #include <sys/mman.h>
     43 
     44 #include <sys/atomics.h>
     45 
     46 /* special private C library header - see Android.mk */
     47 #include <bionic_tls.h>
     48 
     49 #include "linker.h"
     50 #include "linker_debug.h"
     51 #include "linker_environ.h"
     52 #include "linker_format.h"
     53 
     54 #define ALLOW_SYMBOLS_FROM_MAIN 1
     55 #define SO_MAX 128
     56 
     57 /* Assume average path length of 64 and max 8 paths */
     58 #define LDPATH_BUFSIZE 512
     59 #define LDPATH_MAX 8
     60 
     61 #define LDPRELOAD_BUFSIZE 512
     62 #define LDPRELOAD_MAX 8
     63 
     64 /* >>> IMPORTANT NOTE - READ ME BEFORE MODIFYING <<<
     65  *
     66  * Do NOT use malloc() and friends or pthread_*() code here.
     67  * Don't use printf() either; it's caused mysterious memory
     68  * corruption in the past.
     69  * The linker runs before we bring up libc and it's easiest
     70  * to make sure it does not depend on any complex libc features
     71  *
     72  * open issues / todo:
     73  *
     74  * - are we doing everything we should for ARM_COPY relocations?
     75  * - cleaner error reporting
     76  * - after linking, set as much stuff as possible to READONLY
     77  *   and NOEXEC
     78  * - linker hardcodes PAGE_SIZE and PAGE_MASK because the kernel
     79  *   headers provide versions that are negative...
     80  * - allocate space for soinfo structs dynamically instead of
     81  *   having a hard limit (64)
     82 */
     83 
     84 
     85 static int link_image(soinfo *si, unsigned wr_offset);
     86 
     87 static int socount = 0;
     88 static soinfo sopool[SO_MAX];
     89 static soinfo *freelist = NULL;
     90 static soinfo *solist = &libdl_info;
     91 static soinfo *sonext = &libdl_info;
     92 #if ALLOW_SYMBOLS_FROM_MAIN
     93 static soinfo *somain; /* main process, always the one after libdl_info */
     94 #endif
     95 
     96 
     97 static inline int validate_soinfo(soinfo *si)
     98 {
     99     return (si >= sopool && si < sopool + SO_MAX) ||
    100         si == &libdl_info;
    101 }
    102 
    103 static char ldpaths_buf[LDPATH_BUFSIZE];
    104 static const char *ldpaths[LDPATH_MAX + 1];
    105 
    106 static char ldpreloads_buf[LDPRELOAD_BUFSIZE];
    107 static const char *ldpreload_names[LDPRELOAD_MAX + 1];
    108 
    109 static soinfo *preloads[LDPRELOAD_MAX + 1];
    110 
    111 #if LINKER_DEBUG
    112 int debug_verbosity;
    113 #endif
    114 
    115 static int pid;
    116 
    117 /* This boolean is set if the program being loaded is setuid */
    118 static int program_is_setuid;
    119 
    120 #if STATS
    121 struct _link_stats linker_stats;
    122 #endif
    123 
    124 #if COUNT_PAGES
    125 unsigned bitmask[4096];
    126 #endif
    127 
    128 #ifndef PT_ARM_EXIDX
    129 #define PT_ARM_EXIDX    0x70000001      /* .ARM.exidx segment */
    130 #endif
    131 
    132 #define HOODLUM(name, ret, ...)                                               \
    133     ret name __VA_ARGS__                                                      \
    134     {                                                                         \
    135         char errstr[] = "ERROR: " #name " called from the dynamic linker!\n"; \
    136         write(2, errstr, sizeof(errstr));                                     \
    137         abort();                                                              \
    138     }
    139 HOODLUM(malloc, void *, (size_t size));
    140 HOODLUM(free, void, (void *ptr));
    141 HOODLUM(realloc, void *, (void *ptr, size_t size));
    142 HOODLUM(calloc, void *, (size_t cnt, size_t size));
    143 
    144 static char tmp_err_buf[768];
    145 static char __linker_dl_err_buf[768];
    146 #define DL_ERR(fmt, x...)                                                     \
    147     do {                                                                      \
    148         format_buffer(__linker_dl_err_buf, sizeof(__linker_dl_err_buf),            \
    149                  "%s[%d]: " fmt, __func__, __LINE__, ##x);                    \
    150         ERROR(fmt "\n", ##x);                                                      \
    151     } while(0)
    152 
    153 const char *linker_get_error(void)
    154 {
    155     return (const char *)&__linker_dl_err_buf[0];
    156 }
    157 
    158 /*
    159  * This function is an empty stub where GDB locates a breakpoint to get notified
    160  * about linker activity.
    161  */
    162 extern void __attribute__((noinline)) __attribute__((visibility("default"))) rtld_db_dlactivity(void);
    163 
    164 static struct r_debug _r_debug = {1, NULL, &rtld_db_dlactivity,
    165                                   RT_CONSISTENT, 0};
    166 static struct link_map *r_debug_tail = 0;
    167 
    168 static pthread_mutex_t _r_debug_lock = PTHREAD_MUTEX_INITIALIZER;
    169 
    170 static void insert_soinfo_into_debug_map(soinfo * info)
    171 {
    172     struct link_map * map;
    173 
    174     /* Copy the necessary fields into the debug structure.
    175      */
    176     map = &(info->linkmap);
    177     map->l_addr = info->base;
    178     map->l_name = (char*) info->name;
    179     map->l_ld = (uintptr_t)info->dynamic;
    180 
    181     /* Stick the new library at the end of the list.
    182      * gdb tends to care more about libc than it does
    183      * about leaf libraries, and ordering it this way
    184      * reduces the back-and-forth over the wire.
    185      */
    186     if (r_debug_tail) {
    187         r_debug_tail->l_next = map;
    188         map->l_prev = r_debug_tail;
    189         map->l_next = 0;
    190     } else {
    191         _r_debug.r_map = map;
    192         map->l_prev = 0;
    193         map->l_next = 0;
    194     }
    195     r_debug_tail = map;
    196 }
    197 
    198 static void remove_soinfo_from_debug_map(soinfo * info)
    199 {
    200     struct link_map * map = &(info->linkmap);
    201 
    202     if (r_debug_tail == map)
    203         r_debug_tail = map->l_prev;
    204 
    205     if (map->l_prev) map->l_prev->l_next = map->l_next;
    206     if (map->l_next) map->l_next->l_prev = map->l_prev;
    207 }
    208 
    209 void notify_gdb_of_load(soinfo * info)
    210 {
    211     if (info->flags & FLAG_EXE) {
    212         // GDB already knows about the main executable
    213         return;
    214     }
    215 
    216     pthread_mutex_lock(&_r_debug_lock);
    217 
    218     _r_debug.r_state = RT_ADD;
    219     rtld_db_dlactivity();
    220 
    221     insert_soinfo_into_debug_map(info);
    222 
    223     _r_debug.r_state = RT_CONSISTENT;
    224     rtld_db_dlactivity();
    225 
    226     pthread_mutex_unlock(&_r_debug_lock);
    227 }
    228 
    229 void notify_gdb_of_unload(soinfo * info)
    230 {
    231     if (info->flags & FLAG_EXE) {
    232         // GDB already knows about the main executable
    233         return;
    234     }
    235 
    236     pthread_mutex_lock(&_r_debug_lock);
    237 
    238     _r_debug.r_state = RT_DELETE;
    239     rtld_db_dlactivity();
    240 
    241     remove_soinfo_from_debug_map(info);
    242 
    243     _r_debug.r_state = RT_CONSISTENT;
    244     rtld_db_dlactivity();
    245 
    246     pthread_mutex_unlock(&_r_debug_lock);
    247 }
    248 
    249 void notify_gdb_of_libraries()
    250 {
    251     _r_debug.r_state = RT_ADD;
    252     rtld_db_dlactivity();
    253     _r_debug.r_state = RT_CONSISTENT;
    254     rtld_db_dlactivity();
    255 }
    256 
    257 static soinfo *alloc_info(const char *name)
    258 {
    259     soinfo *si;
    260 
    261     if(strlen(name) >= SOINFO_NAME_LEN) {
    262         DL_ERR("%5d library name %s too long", pid, name);
    263         return NULL;
    264     }
    265 
    266     /* The freelist is populated when we call free_info(), which in turn is
    267        done only by dlclose(), which is not likely to be used.
    268     */
    269     if (!freelist) {
    270         if(socount == SO_MAX) {
    271             DL_ERR("%5d too many libraries when loading %s", pid, name);
    272             return NULL;
    273         }
    274         freelist = sopool + socount++;
    275         freelist->next = NULL;
    276     }
    277 
    278     si = freelist;
    279     freelist = freelist->next;
    280 
    281     /* Make sure we get a clean block of soinfo */
    282     memset(si, 0, sizeof(soinfo));
    283     strlcpy((char*) si->name, name, sizeof(si->name));
    284     sonext->next = si;
    285     si->next = NULL;
    286     si->refcount = 0;
    287     sonext = si;
    288 
    289     TRACE("%5d name %s: allocated soinfo @ %p\n", pid, name, si);
    290     return si;
    291 }
    292 
    293 static void free_info(soinfo *si)
    294 {
    295     soinfo *prev = NULL, *trav;
    296 
    297     TRACE("%5d name %s: freeing soinfo @ %p\n", pid, si->name, si);
    298 
    299     for(trav = solist; trav != NULL; trav = trav->next){
    300         if (trav == si)
    301             break;
    302         prev = trav;
    303     }
    304     if (trav == NULL) {
    305         /* si was not ni solist */
    306         DL_ERR("%5d name %s is not in solist!", pid, si->name);
    307         return;
    308     }
    309 
    310     /* prev will never be NULL, because the first entry in solist is
    311        always the static libdl_info.
    312     */
    313     prev->next = si->next;
    314     if (si == sonext) sonext = prev;
    315     si->next = freelist;
    316     freelist = si;
    317 }
    318 
    319 const char *addr_to_name(unsigned addr)
    320 {
    321     soinfo *si;
    322 
    323     for(si = solist; si != 0; si = si->next){
    324         if((addr >= si->base) && (addr < (si->base + si->size))) {
    325             return si->name;
    326         }
    327     }
    328 
    329     return "";
    330 }
    331 
    332 /* For a given PC, find the .so that it belongs to.
    333  * Returns the base address of the .ARM.exidx section
    334  * for that .so, and the number of 8-byte entries
    335  * in that section (via *pcount).
    336  *
    337  * Intended to be called by libc's __gnu_Unwind_Find_exidx().
    338  *
    339  * This function is exposed via dlfcn.c and libdl.so.
    340  */
    341 #ifdef ANDROID_ARM_LINKER
    342 _Unwind_Ptr dl_unwind_find_exidx(_Unwind_Ptr pc, int *pcount)
    343 {
    344     soinfo *si;
    345     unsigned addr = (unsigned)pc;
    346 
    347     for (si = solist; si != 0; si = si->next){
    348         if ((addr >= si->base) && (addr < (si->base + si->size))) {
    349             *pcount = si->ARM_exidx_count;
    350             return (_Unwind_Ptr)(si->base + (unsigned long)si->ARM_exidx);
    351         }
    352     }
    353    *pcount = 0;
    354     return NULL;
    355 }
    356 #elif defined(ANDROID_X86_LINKER)
    357 /* Here, we only have to provide a callback to iterate across all the
    358  * loaded libraries. gcc_eh does the rest. */
    359 int
    360 dl_iterate_phdr(int (*cb)(struct dl_phdr_info *info, size_t size, void *data),
    361                 void *data)
    362 {
    363     soinfo *si;
    364     struct dl_phdr_info dl_info;
    365     int rv = 0;
    366 
    367     for (si = solist; si != NULL; si = si->next) {
    368         dl_info.dlpi_addr = si->linkmap.l_addr;
    369         dl_info.dlpi_name = si->linkmap.l_name;
    370         dl_info.dlpi_phdr = si->phdr;
    371         dl_info.dlpi_phnum = si->phnum;
    372         rv = cb(&dl_info, sizeof (struct dl_phdr_info), data);
    373         if (rv != 0)
    374             break;
    375     }
    376     return rv;
    377 }
    378 #endif
    379 
    380 static Elf32_Sym *_elf_lookup(soinfo *si, unsigned hash, const char *name)
    381 {
    382     Elf32_Sym *s;
    383     Elf32_Sym *symtab = si->symtab;
    384     const char *strtab = si->strtab;
    385     unsigned n;
    386 
    387     TRACE_TYPE(LOOKUP, "%5d SEARCH %s in %s@0x%08x %08x %d\n", pid,
    388                name, si->name, si->base, hash, hash % si->nbucket);
    389     n = hash % si->nbucket;
    390 
    391     for(n = si->bucket[hash % si->nbucket]; n != 0; n = si->chain[n]){
    392         s = symtab + n;
    393         if(strcmp(strtab + s->st_name, name)) continue;
    394 
    395             /* only concern ourselves with global and weak symbol definitions */
    396         switch(ELF32_ST_BIND(s->st_info)){
    397         case STB_GLOBAL:
    398         case STB_WEAK:
    399                 /* no section == undefined */
    400             if(s->st_shndx == 0) continue;
    401 
    402             TRACE_TYPE(LOOKUP, "%5d FOUND %s in %s (%08x) %d\n", pid,
    403                        name, si->name, s->st_value, s->st_size);
    404             return s;
    405         }
    406     }
    407 
    408     return NULL;
    409 }
    410 
    411 static unsigned elfhash(const char *_name)
    412 {
    413     const unsigned char *name = (const unsigned char *) _name;
    414     unsigned h = 0, g;
    415 
    416     while(*name) {
    417         h = (h << 4) + *name++;
    418         g = h & 0xf0000000;
    419         h ^= g;
    420         h ^= g >> 24;
    421     }
    422     return h;
    423 }
    424 
    425 static Elf32_Sym *
    426 _do_lookup(soinfo *si, const char *name, unsigned *base)
    427 {
    428     unsigned elf_hash = elfhash(name);
    429     Elf32_Sym *s;
    430     unsigned *d;
    431     soinfo *lsi = si;
    432     int i;
    433 
    434     /* Look for symbols in the local scope (the object who is
    435      * searching). This happens with C++ templates on i386 for some
    436      * reason.
    437      *
    438      * Notes on weak symbols:
    439      * The ELF specs are ambigious about treatment of weak definitions in
    440      * dynamic linking.  Some systems return the first definition found
    441      * and some the first non-weak definition.   This is system dependent.
    442      * Here we return the first definition found for simplicity.  */
    443 
    444     s = _elf_lookup(si, elf_hash, name);
    445     if(s != NULL)
    446         goto done;
    447 
    448     /* Next, look for it in the preloads list */
    449     for(i = 0; preloads[i] != NULL; i++) {
    450         lsi = preloads[i];
    451         s = _elf_lookup(lsi, elf_hash, name);
    452         if(s != NULL)
    453             goto done;
    454     }
    455 
    456     for(d = si->dynamic; *d; d += 2) {
    457         if(d[0] == DT_NEEDED){
    458             lsi = (soinfo *)d[1];
    459             if (!validate_soinfo(lsi)) {
    460                 DL_ERR("%5d bad DT_NEEDED pointer in %s",
    461                        pid, si->name);
    462                 return NULL;
    463             }
    464 
    465             DEBUG("%5d %s: looking up %s in %s\n",
    466                   pid, si->name, name, lsi->name);
    467             s = _elf_lookup(lsi, elf_hash, name);
    468             if ((s != NULL) && (s->st_shndx != SHN_UNDEF))
    469                 goto done;
    470         }
    471     }
    472 
    473 #if ALLOW_SYMBOLS_FROM_MAIN
    474     /* If we are resolving relocations while dlopen()ing a library, it's OK for
    475      * the library to resolve a symbol that's defined in the executable itself,
    476      * although this is rare and is generally a bad idea.
    477      */
    478     if (somain) {
    479         lsi = somain;
    480         DEBUG("%5d %s: looking up %s in executable %s\n",
    481               pid, si->name, name, lsi->name);
    482         s = _elf_lookup(lsi, elf_hash, name);
    483     }
    484 #endif
    485 
    486 done:
    487     if(s != NULL) {
    488         TRACE_TYPE(LOOKUP, "%5d si %s sym %s s->st_value = 0x%08x, "
    489                    "found in %s, base = 0x%08x\n",
    490                    pid, si->name, name, s->st_value, lsi->name, lsi->base);
    491         *base = lsi->base;
    492         return s;
    493     }
    494 
    495     return NULL;
    496 }
    497 
    498 /* This is used by dl_sym().  It performs symbol lookup only within the
    499    specified soinfo object and not in any of its dependencies.
    500  */
    501 Elf32_Sym *lookup_in_library(soinfo *si, const char *name)
    502 {
    503     return _elf_lookup(si, elfhash(name), name);
    504 }
    505 
    506 /* This is used by dl_sym().  It performs a global symbol lookup.
    507  */
    508 Elf32_Sym *lookup(const char *name, soinfo **found, soinfo *start)
    509 {
    510     unsigned elf_hash = elfhash(name);
    511     Elf32_Sym *s = NULL;
    512     soinfo *si;
    513 
    514     if(start == NULL) {
    515         start = solist;
    516     }
    517 
    518     for(si = start; (s == NULL) && (si != NULL); si = si->next)
    519     {
    520         if(si->flags & FLAG_ERROR)
    521             continue;
    522         s = _elf_lookup(si, elf_hash, name);
    523         if (s != NULL) {
    524             *found = si;
    525             break;
    526         }
    527     }
    528 
    529     if(s != NULL) {
    530         TRACE_TYPE(LOOKUP, "%5d %s s->st_value = 0x%08x, "
    531                    "si->base = 0x%08x\n", pid, name, s->st_value, si->base);
    532         return s;
    533     }
    534 
    535     return NULL;
    536 }
    537 
    538 soinfo *find_containing_library(const void *addr)
    539 {
    540     soinfo *si;
    541 
    542     for(si = solist; si != NULL; si = si->next)
    543     {
    544         if((unsigned)addr >= si->base && (unsigned)addr - si->base < si->size) {
    545             return si;
    546         }
    547     }
    548 
    549     return NULL;
    550 }
    551 
    552 Elf32_Sym *find_containing_symbol(const void *addr, soinfo *si)
    553 {
    554     unsigned int i;
    555     unsigned soaddr = (unsigned)addr - si->base;
    556 
    557     /* Search the library's symbol table for any defined symbol which
    558      * contains this address */
    559     for(i=0; i<si->nchain; i++) {
    560         Elf32_Sym *sym = &si->symtab[i];
    561 
    562         if(sym->st_shndx != SHN_UNDEF &&
    563            soaddr >= sym->st_value &&
    564            soaddr < sym->st_value + sym->st_size) {
    565             return sym;
    566         }
    567     }
    568 
    569     return NULL;
    570 }
    571 
    572 #if 0
    573 static void dump(soinfo *si)
    574 {
    575     Elf32_Sym *s = si->symtab;
    576     unsigned n;
    577 
    578     for(n = 0; n < si->nchain; n++) {
    579         TRACE("%5d %04d> %08x: %02x %04x %08x %08x %s\n", pid, n, s,
    580                s->st_info, s->st_shndx, s->st_value, s->st_size,
    581                si->strtab + s->st_name);
    582         s++;
    583     }
    584 }
    585 #endif
    586 
    587 static const char *sopaths[] = {
    588     "/vendor/lib",
    589     "/system/lib",
    590     0
    591 };
    592 
    593 static int _open_lib(const char *name)
    594 {
    595     int fd;
    596     struct stat filestat;
    597 
    598     if ((stat(name, &filestat) >= 0) && S_ISREG(filestat.st_mode)) {
    599         if ((fd = open(name, O_RDONLY)) >= 0)
    600             return fd;
    601     }
    602 
    603     return -1;
    604 }
    605 
    606 static int open_library(const char *name)
    607 {
    608     int fd;
    609     char buf[512];
    610     const char **path;
    611     int n;
    612 
    613     TRACE("[ %5d opening %s ]\n", pid, name);
    614 
    615     if(name == 0) return -1;
    616     if(strlen(name) > 256) return -1;
    617 
    618     if ((name[0] == '/') && ((fd = _open_lib(name)) >= 0))
    619         return fd;
    620 
    621     for (path = ldpaths; *path; path++) {
    622         n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
    623         if (n < 0 || n >= (int)sizeof(buf)) {
    624             WARN("Ignoring very long library path: %s/%s\n", *path, name);
    625             continue;
    626         }
    627         if ((fd = _open_lib(buf)) >= 0)
    628             return fd;
    629     }
    630     for (path = sopaths; *path; path++) {
    631         n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
    632         if (n < 0 || n >= (int)sizeof(buf)) {
    633             WARN("Ignoring very long library path: %s/%s\n", *path, name);
    634             continue;
    635         }
    636         if ((fd = _open_lib(buf)) >= 0)
    637             return fd;
    638     }
    639 
    640     return -1;
    641 }
    642 
    643 /* temporary space for holding the first page of the shared lib
    644  * which contains the elf header (with the pht). */
    645 static unsigned char __header[PAGE_SIZE];
    646 
    647 typedef struct {
    648     long mmap_addr;
    649     char tag[4]; /* 'P', 'R', 'E', ' ' */
    650 } prelink_info_t;
    651 
    652 /* Returns the requested base address if the library is prelinked,
    653  * and 0 otherwise.  */
    654 static unsigned long
    655 is_prelinked(int fd, const char *name)
    656 {
    657     off_t sz;
    658     prelink_info_t info;
    659 
    660     sz = lseek(fd, -sizeof(prelink_info_t), SEEK_END);
    661     if (sz < 0) {
    662         DL_ERR("lseek() failed!");
    663         return 0;
    664     }
    665 
    666     if (read(fd, &info, sizeof(info)) != sizeof(info)) {
    667         WARN("Could not read prelink_info_t structure for `%s`\n", name);
    668         return 0;
    669     }
    670 
    671     if (strncmp(info.tag, "PRE ", 4)) {
    672         WARN("`%s` is not a prelinked library\n", name);
    673         return 0;
    674     }
    675 
    676     return (unsigned long)info.mmap_addr;
    677 }
    678 
    679 /* verify_elf_object
    680  *      Verifies if the object @ base is a valid ELF object
    681  *
    682  * Args:
    683  *
    684  * Returns:
    685  *       0 on success
    686  *      -1 if no valid ELF object is found @ base.
    687  */
    688 static int
    689 verify_elf_object(void *base, const char *name)
    690 {
    691     Elf32_Ehdr *hdr = (Elf32_Ehdr *) base;
    692 
    693     if (hdr->e_ident[EI_MAG0] != ELFMAG0) return -1;
    694     if (hdr->e_ident[EI_MAG1] != ELFMAG1) return -1;
    695     if (hdr->e_ident[EI_MAG2] != ELFMAG2) return -1;
    696     if (hdr->e_ident[EI_MAG3] != ELFMAG3) return -1;
    697 
    698     /* TODO: Should we verify anything else in the header? */
    699 #ifdef ANDROID_ARM_LINKER
    700     if (hdr->e_machine != EM_ARM) return -1;
    701 #elif defined(ANDROID_X86_LINKER)
    702     if (hdr->e_machine != EM_386) return -1;
    703 #endif
    704     return 0;
    705 }
    706 
    707 
    708 /* get_lib_extents
    709  *      Retrieves the base (*base) address where the ELF object should be
    710  *      mapped and its overall memory size (*total_sz).
    711  *
    712  * Args:
    713  *      fd: Opened file descriptor for the library
    714  *      name: The name of the library
    715  *      _hdr: Pointer to the header page of the library
    716  *      total_sz: Total size of the memory that should be allocated for
    717  *                this library
    718  *
    719  * Returns:
    720  *      -1 if there was an error while trying to get the lib extents.
    721  *         The possible reasons are:
    722  *             - Could not determine if the library was prelinked.
    723  *             - The library provided is not a valid ELF object
    724  *       0 if the library did not request a specific base offset (normal
    725  *         for non-prelinked libs)
    726  *     > 0 if the library requests a specific address to be mapped to.
    727  *         This indicates a pre-linked library.
    728  */
    729 static unsigned
    730 get_lib_extents(int fd, const char *name, void *__hdr, unsigned *total_sz)
    731 {
    732     unsigned req_base;
    733     unsigned min_vaddr = 0xffffffff;
    734     unsigned max_vaddr = 0;
    735     unsigned char *_hdr = (unsigned char *)__hdr;
    736     Elf32_Ehdr *ehdr = (Elf32_Ehdr *)_hdr;
    737     Elf32_Phdr *phdr;
    738     int cnt;
    739 
    740     TRACE("[ %5d Computing extents for '%s'. ]\n", pid, name);
    741     if (verify_elf_object(_hdr, name) < 0) {
    742         DL_ERR("%5d - %s is not a valid ELF object", pid, name);
    743         return (unsigned)-1;
    744     }
    745 
    746     req_base = (unsigned) is_prelinked(fd, name);
    747     if (req_base == (unsigned)-1)
    748         return -1;
    749     else if (req_base != 0) {
    750         TRACE("[ %5d - Prelinked library '%s' requesting base @ 0x%08x ]\n",
    751               pid, name, req_base);
    752     } else {
    753         TRACE("[ %5d - Non-prelinked library '%s' found. ]\n", pid, name);
    754     }
    755 
    756     phdr = (Elf32_Phdr *)(_hdr + ehdr->e_phoff);
    757 
    758     /* find the min/max p_vaddrs from all the PT_LOAD segments so we can
    759      * get the range. */
    760     for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
    761         if (phdr->p_type == PT_LOAD) {
    762             if ((phdr->p_vaddr + phdr->p_memsz) > max_vaddr)
    763                 max_vaddr = phdr->p_vaddr + phdr->p_memsz;
    764             if (phdr->p_vaddr < min_vaddr)
    765                 min_vaddr = phdr->p_vaddr;
    766         }
    767     }
    768 
    769     if ((min_vaddr == 0xffffffff) && (max_vaddr == 0)) {
    770         DL_ERR("%5d - No loadable segments found in %s.", pid, name);
    771         return (unsigned)-1;
    772     }
    773 
    774     /* truncate min_vaddr down to page boundary */
    775     min_vaddr &= ~PAGE_MASK;
    776 
    777     /* round max_vaddr up to the next page */
    778     max_vaddr = (max_vaddr + PAGE_SIZE - 1) & ~PAGE_MASK;
    779 
    780     *total_sz = (max_vaddr - min_vaddr);
    781     return (unsigned)req_base;
    782 }
    783 
    784 /* reserve_mem_region
    785  *
    786  *     This function reserves a chunk of memory to be used for mapping in
    787  *     a prelinked shared library. We reserve the entire memory region here, and
    788  *     then the rest of the linker will relocate the individual loadable
    789  *     segments into the correct locations within this memory range.
    790  *
    791  * Args:
    792  *     si->base: The requested base of the allocation.
    793  *     si->size: The size of the allocation.
    794  *
    795  * Returns:
    796  *     -1 on failure, and 0 on success.  On success, si->base will contain
    797  *     the virtual address at which the library will be mapped.
    798  */
    799 
    800 static int reserve_mem_region(soinfo *si)
    801 {
    802     void *base = mmap((void *)si->base, si->size, PROT_NONE,
    803                       MAP_FIXED | MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
    804     if (base == MAP_FAILED) {
    805         DL_ERR("%5d can NOT map (%sprelinked) library '%s' at 0x%08x "
    806               "as requested, will try general pool: %d (%s)",
    807               pid, (si->base ? "" : "non-"), si->name, si->base,
    808               errno, strerror(errno));
    809         return -1;
    810     } else if (base != (void *)si->base) {
    811         DL_ERR("OOPS: %5d %sprelinked library '%s' mapped at 0x%08x, "
    812               "not at 0x%08x", pid, (si->base ? "" : "non-"),
    813               si->name, (unsigned)base, si->base);
    814         munmap(base, si->size);
    815         return -1;
    816     }
    817     return 0;
    818 }
    819 
    820 static int alloc_mem_region(soinfo *si)
    821 {
    822     if (si->base) {
    823         /* Attempt to mmap a prelinked library. */
    824         return reserve_mem_region(si);
    825     }
    826 
    827     /* This is not a prelinked library, so we use the kernel's default
    828        allocator.
    829     */
    830 
    831     void *base = mmap(NULL, si->size, PROT_NONE,
    832                       MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
    833     if (base == MAP_FAILED) {
    834         DL_ERR("%5d mmap of library '%s' failed: %d (%s)\n",
    835               pid, si->name,
    836               errno, strerror(errno));
    837         goto err;
    838     }
    839     si->base = (unsigned) base;
    840     PRINT("%5d mapped library '%s' to %08x via kernel allocator.\n",
    841           pid, si->name, si->base);
    842     return 0;
    843 
    844 err:
    845     DL_ERR("OOPS: %5d cannot map library '%s'. no vspace available.",
    846           pid, si->name);
    847     return -1;
    848 }
    849 
    850 #define MAYBE_MAP_FLAG(x,from,to)    (((x) & (from)) ? (to) : 0)
    851 #define PFLAGS_TO_PROT(x)            (MAYBE_MAP_FLAG((x), PF_X, PROT_EXEC) | \
    852                                       MAYBE_MAP_FLAG((x), PF_R, PROT_READ) | \
    853                                       MAYBE_MAP_FLAG((x), PF_W, PROT_WRITE))
    854 /* load_segments
    855  *
    856  *     This function loads all the loadable (PT_LOAD) segments into memory
    857  *     at their appropriate memory offsets off the base address.
    858  *
    859  * Args:
    860  *     fd: Open file descriptor to the library to load.
    861  *     header: Pointer to a header page that contains the ELF header.
    862  *             This is needed since we haven't mapped in the real file yet.
    863  *     si: ptr to soinfo struct describing the shared object.
    864  *
    865  * Returns:
    866  *     0 on success, -1 on failure.
    867  */
    868 static int
    869 load_segments(int fd, void *header, soinfo *si)
    870 {
    871     Elf32_Ehdr *ehdr = (Elf32_Ehdr *)header;
    872     Elf32_Phdr *phdr = (Elf32_Phdr *)((unsigned char *)header + ehdr->e_phoff);
    873     Elf32_Addr base = (Elf32_Addr) si->base;
    874     int cnt;
    875     unsigned len;
    876     Elf32_Addr tmp;
    877     unsigned char *pbase;
    878     unsigned char *extra_base;
    879     unsigned extra_len;
    880     unsigned total_sz = 0;
    881 
    882     si->wrprotect_start = 0xffffffff;
    883     si->wrprotect_end = 0;
    884 
    885     TRACE("[ %5d - Begin loading segments for '%s' @ 0x%08x ]\n",
    886           pid, si->name, (unsigned)si->base);
    887     /* Now go through all the PT_LOAD segments and map them into memory
    888      * at the appropriate locations. */
    889     for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
    890         if (phdr->p_type == PT_LOAD) {
    891             DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
    892             /* we want to map in the segment on a page boundary */
    893             tmp = base + (phdr->p_vaddr & (~PAGE_MASK));
    894             /* add the # of bytes we masked off above to the total length. */
    895             len = phdr->p_filesz + (phdr->p_vaddr & PAGE_MASK);
    896 
    897             TRACE("[ %d - Trying to load segment from '%s' @ 0x%08x "
    898                   "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x ]\n", pid, si->name,
    899                   (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
    900             pbase = mmap((void *)tmp, len, PFLAGS_TO_PROT(phdr->p_flags),
    901                          MAP_PRIVATE | MAP_FIXED, fd,
    902                          phdr->p_offset & (~PAGE_MASK));
    903             if (pbase == MAP_FAILED) {
    904                 DL_ERR("%d failed to map segment from '%s' @ 0x%08x (0x%08x). "
    905                       "p_vaddr=0x%08x p_offset=0x%08x", pid, si->name,
    906                       (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
    907                 goto fail;
    908             }
    909 
    910             /* If 'len' didn't end on page boundary, and it's a writable
    911              * segment, zero-fill the rest. */
    912             if ((len & PAGE_MASK) && (phdr->p_flags & PF_W))
    913                 memset((void *)(pbase + len), 0, PAGE_SIZE - (len & PAGE_MASK));
    914 
    915             /* Check to see if we need to extend the map for this segment to
    916              * cover the diff between filesz and memsz (i.e. for bss).
    917              *
    918              *  base           _+---------------------+  page boundary
    919              *                  .                     .
    920              *                  |                     |
    921              *                  .                     .
    922              *  pbase          _+---------------------+  page boundary
    923              *                  |                     |
    924              *                  .                     .
    925              *  base + p_vaddr _|                     |
    926              *                  . \          \        .
    927              *                  . | filesz   |        .
    928              *  pbase + len    _| /          |        |
    929              *     <0 pad>      .            .        .
    930              *  extra_base     _+------------|--------+  page boundary
    931              *               /  .            .        .
    932              *               |  .            .        .
    933              *               |  +------------|--------+  page boundary
    934              *  extra_len->  |  |            |        |
    935              *               |  .            | memsz  .
    936              *               |  .            |        .
    937              *               \ _|            /        |
    938              *                  .                     .
    939              *                  |                     |
    940              *                 _+---------------------+  page boundary
    941              */
    942             tmp = (Elf32_Addr)(((unsigned)pbase + len + PAGE_SIZE - 1) &
    943                                     (~PAGE_MASK));
    944             if (tmp < (base + phdr->p_vaddr + phdr->p_memsz)) {
    945                 extra_len = base + phdr->p_vaddr + phdr->p_memsz - tmp;
    946                 TRACE("[ %5d - Need to extend segment from '%s' @ 0x%08x "
    947                       "(0x%08x) ]\n", pid, si->name, (unsigned)tmp, extra_len);
    948                 /* map in the extra page(s) as anonymous into the range.
    949                  * This is probably not necessary as we already mapped in
    950                  * the entire region previously, but we just want to be
    951                  * sure. This will also set the right flags on the region
    952                  * (though we can probably accomplish the same thing with
    953                  * mprotect).
    954                  */
    955                 extra_base = mmap((void *)tmp, extra_len,
    956                                   PFLAGS_TO_PROT(phdr->p_flags),
    957                                   MAP_PRIVATE | MAP_FIXED | MAP_ANONYMOUS,
    958                                   -1, 0);
    959                 if (extra_base == MAP_FAILED) {
    960                     DL_ERR("[ %5d - failed to extend segment from '%s' @ 0x%08x"
    961                            " (0x%08x) ]", pid, si->name, (unsigned)tmp,
    962                           extra_len);
    963                     goto fail;
    964                 }
    965                 /* TODO: Check if we need to memset-0 this region.
    966                  * Anonymous mappings are zero-filled copy-on-writes, so we
    967                  * shouldn't need to. */
    968                 TRACE("[ %5d - Segment from '%s' extended @ 0x%08x "
    969                       "(0x%08x)\n", pid, si->name, (unsigned)extra_base,
    970                       extra_len);
    971             }
    972             /* set the len here to show the full extent of the segment we
    973              * just loaded, mostly for debugging */
    974             len = (((unsigned)base + phdr->p_vaddr + phdr->p_memsz +
    975                     PAGE_SIZE - 1) & (~PAGE_MASK)) - (unsigned)pbase;
    976             TRACE("[ %5d - Successfully loaded segment from '%s' @ 0x%08x "
    977                   "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x\n", pid, si->name,
    978                   (unsigned)pbase, len, phdr->p_vaddr, phdr->p_offset);
    979             total_sz += len;
    980             /* Make the section writable just in case we'll have to write to
    981              * it during relocation (i.e. text segment). However, we will
    982              * remember what range of addresses should be write protected.
    983              *
    984              */
    985             if (!(phdr->p_flags & PF_W)) {
    986                 if ((unsigned)pbase < si->wrprotect_start)
    987                     si->wrprotect_start = (unsigned)pbase;
    988                 if (((unsigned)pbase + len) > si->wrprotect_end)
    989                     si->wrprotect_end = (unsigned)pbase + len;
    990                 mprotect(pbase, len,
    991                          PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
    992             }
    993         } else if (phdr->p_type == PT_DYNAMIC) {
    994             DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
    995             /* this segment contains the dynamic linking information */
    996             si->dynamic = (unsigned *)(base + phdr->p_vaddr);
    997         } else if (phdr->p_type == PT_GNU_RELRO) {
    998             if ((phdr->p_vaddr >= si->size)
    999                     || ((phdr->p_vaddr + phdr->p_memsz) > si->size)
   1000                     || ((base + phdr->p_vaddr + phdr->p_memsz) < base)) {
   1001                 DL_ERR("%d invalid GNU_RELRO in '%s' "
   1002                        "p_vaddr=0x%08x p_memsz=0x%08x", pid, si->name,
   1003                        phdr->p_vaddr, phdr->p_memsz);
   1004                 goto fail;
   1005             }
   1006             si->gnu_relro_start = (Elf32_Addr) (base + phdr->p_vaddr);
   1007             si->gnu_relro_len = (unsigned) phdr->p_memsz;
   1008         } else {
   1009 #ifdef ANDROID_ARM_LINKER
   1010             if (phdr->p_type == PT_ARM_EXIDX) {
   1011                 DEBUG_DUMP_PHDR(phdr, "PT_ARM_EXIDX", pid);
   1012                 /* exidx entries (used for stack unwinding) are 8 bytes each.
   1013                  */
   1014                 si->ARM_exidx = (unsigned *)phdr->p_vaddr;
   1015                 si->ARM_exidx_count = phdr->p_memsz / 8;
   1016             }
   1017 #endif
   1018         }
   1019 
   1020     }
   1021 
   1022     /* Sanity check */
   1023     if (total_sz > si->size) {
   1024         DL_ERR("%5d - Total length (0x%08x) of mapped segments from '%s' is "
   1025               "greater than what was allocated (0x%08x). THIS IS BAD!",
   1026               pid, total_sz, si->name, si->size);
   1027         goto fail;
   1028     }
   1029 
   1030     TRACE("[ %5d - Finish loading segments for '%s' @ 0x%08x. "
   1031           "Total memory footprint: 0x%08x bytes ]\n", pid, si->name,
   1032           (unsigned)si->base, si->size);
   1033     return 0;
   1034 
   1035 fail:
   1036     /* We can just blindly unmap the entire region even though some things
   1037      * were mapped in originally with anonymous and others could have been
   1038      * been mapped in from the file before we failed. The kernel will unmap
   1039      * all the pages in the range, irrespective of how they got there.
   1040      */
   1041     munmap((void *)si->base, si->size);
   1042     si->flags |= FLAG_ERROR;
   1043     return -1;
   1044 }
   1045 
   1046 /* TODO: Implement this to take care of the fact that Android ARM
   1047  * ELF objects shove everything into a single loadable segment that has the
   1048  * write bit set. wr_offset is then used to set non-(data|bss) pages to be
   1049  * non-writable.
   1050  */
   1051 #if 0
   1052 static unsigned
   1053 get_wr_offset(int fd, const char *name, Elf32_Ehdr *ehdr)
   1054 {
   1055     Elf32_Shdr *shdr_start;
   1056     Elf32_Shdr *shdr;
   1057     int shdr_sz = ehdr->e_shnum * sizeof(Elf32_Shdr);
   1058     int cnt;
   1059     unsigned wr_offset = 0xffffffff;
   1060 
   1061     shdr_start = mmap(0, shdr_sz, PROT_READ, MAP_PRIVATE, fd,
   1062                       ehdr->e_shoff & (~PAGE_MASK));
   1063     if (shdr_start == MAP_FAILED) {
   1064         WARN("%5d - Could not read section header info from '%s'. Will not "
   1065              "not be able to determine write-protect offset.\n", pid, name);
   1066         return (unsigned)-1;
   1067     }
   1068 
   1069     for(cnt = 0, shdr = shdr_start; cnt < ehdr->e_shnum; ++cnt, ++shdr) {
   1070         if ((shdr->sh_type != SHT_NULL) && (shdr->sh_flags & SHF_WRITE) &&
   1071             (shdr->sh_addr < wr_offset)) {
   1072             wr_offset = shdr->sh_addr;
   1073         }
   1074     }
   1075 
   1076     munmap(shdr_start, shdr_sz);
   1077     return wr_offset;
   1078 }
   1079 #endif
   1080 
   1081 static soinfo *
   1082 load_library(const char *name)
   1083 {
   1084     int fd = open_library(name);
   1085     int cnt;
   1086     unsigned ext_sz;
   1087     unsigned req_base;
   1088     const char *bname;
   1089     soinfo *si = NULL;
   1090     Elf32_Ehdr *hdr;
   1091 
   1092     if(fd == -1) {
   1093         DL_ERR("Library '%s' not found", name);
   1094         return NULL;
   1095     }
   1096 
   1097     /* We have to read the ELF header to figure out what to do with this image
   1098      */
   1099     if (lseek(fd, 0, SEEK_SET) < 0) {
   1100         DL_ERR("lseek() failed!");
   1101         goto fail;
   1102     }
   1103 
   1104     if ((cnt = read(fd, &__header[0], PAGE_SIZE)) < 0) {
   1105         DL_ERR("read() failed!");
   1106         goto fail;
   1107     }
   1108 
   1109     /* Parse the ELF header and get the size of the memory footprint for
   1110      * the library */
   1111     req_base = get_lib_extents(fd, name, &__header[0], &ext_sz);
   1112     if (req_base == (unsigned)-1)
   1113         goto fail;
   1114     TRACE("[ %5d - '%s' (%s) wants base=0x%08x sz=0x%08x ]\n", pid, name,
   1115           (req_base ? "prelinked" : "not pre-linked"), req_base, ext_sz);
   1116 
   1117     /* Now configure the soinfo struct where we'll store all of our data
   1118      * for the ELF object. If the loading fails, we waste the entry, but
   1119      * same thing would happen if we failed during linking. Configuring the
   1120      * soinfo struct here is a lot more convenient.
   1121      */
   1122     bname = strrchr(name, '/');
   1123     si = alloc_info(bname ? bname + 1 : name);
   1124     if (si == NULL)
   1125         goto fail;
   1126 
   1127     /* Carve out a chunk of memory where we will map in the individual
   1128      * segments */
   1129     si->base = req_base;
   1130     si->size = ext_sz;
   1131     si->flags = 0;
   1132     si->entry = 0;
   1133     si->dynamic = (unsigned *)-1;
   1134     if (alloc_mem_region(si) < 0)
   1135         goto fail;
   1136 
   1137     TRACE("[ %5d allocated memory for %s @ %p (0x%08x) ]\n",
   1138           pid, name, (void *)si->base, (unsigned) ext_sz);
   1139 
   1140     /* Now actually load the library's segments into right places in memory */
   1141     if (load_segments(fd, &__header[0], si) < 0) {
   1142         goto fail;
   1143     }
   1144 
   1145     /* this might not be right. Technically, we don't even need this info
   1146      * once we go through 'load_segments'. */
   1147     hdr = (Elf32_Ehdr *)si->base;
   1148     si->phdr = (Elf32_Phdr *)((unsigned char *)si->base + hdr->e_phoff);
   1149     si->phnum = hdr->e_phnum;
   1150     /**/
   1151 
   1152     close(fd);
   1153     return si;
   1154 
   1155 fail:
   1156     if (si) free_info(si);
   1157     close(fd);
   1158     return NULL;
   1159 }
   1160 
   1161 static soinfo *
   1162 init_library(soinfo *si)
   1163 {
   1164     unsigned wr_offset = 0xffffffff;
   1165 
   1166     /* At this point we know that whatever is loaded @ base is a valid ELF
   1167      * shared library whose segments are properly mapped in. */
   1168     TRACE("[ %5d init_library base=0x%08x sz=0x%08x name='%s') ]\n",
   1169           pid, si->base, si->size, si->name);
   1170 
   1171     if(link_image(si, wr_offset)) {
   1172             /* We failed to link.  However, we can only restore libbase
   1173             ** if no additional libraries have moved it since we updated it.
   1174             */
   1175         munmap((void *)si->base, si->size);
   1176         return NULL;
   1177     }
   1178 
   1179     return si;
   1180 }
   1181 
   1182 soinfo *find_library(const char *name)
   1183 {
   1184     soinfo *si;
   1185     const char *bname;
   1186 
   1187 #if ALLOW_SYMBOLS_FROM_MAIN
   1188     if (name == NULL)
   1189         return somain;
   1190 #else
   1191     if (name == NULL)
   1192         return NULL;
   1193 #endif
   1194 
   1195     bname = strrchr(name, '/');
   1196     bname = bname ? bname + 1 : name;
   1197 
   1198     for(si = solist; si != 0; si = si->next){
   1199         if(!strcmp(bname, si->name)) {
   1200             if(si->flags & FLAG_ERROR) {
   1201                 DL_ERR("%5d '%s' failed to load previously", pid, bname);
   1202                 return NULL;
   1203             }
   1204             if(si->flags & FLAG_LINKED) return si;
   1205             DL_ERR("OOPS: %5d recursive link to '%s'", pid, si->name);
   1206             return NULL;
   1207         }
   1208     }
   1209 
   1210     TRACE("[ %5d '%s' has not been loaded yet.  Locating...]\n", pid, name);
   1211     si = load_library(name);
   1212     if(si == NULL)
   1213         return NULL;
   1214     return init_library(si);
   1215 }
   1216 
   1217 /* TODO:
   1218  *   notify gdb of unload
   1219  *   for non-prelinked libraries, find a way to decrement libbase
   1220  */
   1221 static void call_destructors(soinfo *si);
   1222 unsigned unload_library(soinfo *si)
   1223 {
   1224     unsigned *d;
   1225     if (si->refcount == 1) {
   1226         TRACE("%5d unloading '%s'\n", pid, si->name);
   1227         call_destructors(si);
   1228 
   1229         /*
   1230          * Make sure that we undo the PT_GNU_RELRO protections we added
   1231          * in link_image. This is needed to undo the DT_NEEDED hack below.
   1232          */
   1233         if ((si->gnu_relro_start != 0) && (si->gnu_relro_len != 0)) {
   1234             Elf32_Addr start = (si->gnu_relro_start & ~PAGE_MASK);
   1235             unsigned len = (si->gnu_relro_start - start) + si->gnu_relro_len;
   1236             if (mprotect((void *) start, len, PROT_READ | PROT_WRITE) < 0)
   1237                 DL_ERR("%5d %s: could not undo GNU_RELRO protections. "
   1238                        "Expect a crash soon. errno=%d (%s)",
   1239                        pid, si->name, errno, strerror(errno));
   1240 
   1241         }
   1242 
   1243         for(d = si->dynamic; *d; d += 2) {
   1244             if(d[0] == DT_NEEDED){
   1245                 soinfo *lsi = (soinfo *)d[1];
   1246 
   1247                 // The next line will segfault if the we don't undo the
   1248                 // PT_GNU_RELRO protections (see comments above and in
   1249                 // link_image().
   1250                 d[1] = 0;
   1251 
   1252                 if (validate_soinfo(lsi)) {
   1253                     TRACE("%5d %s needs to unload %s\n", pid,
   1254                           si->name, lsi->name);
   1255                     unload_library(lsi);
   1256                 }
   1257                 else
   1258                     DL_ERR("%5d %s: could not unload dependent library",
   1259                            pid, si->name);
   1260             }
   1261         }
   1262 
   1263         munmap((char *)si->base, si->size);
   1264         notify_gdb_of_unload(si);
   1265         free_info(si);
   1266         si->refcount = 0;
   1267     }
   1268     else {
   1269         si->refcount--;
   1270         PRINT("%5d not unloading '%s', decrementing refcount to %d\n",
   1271               pid, si->name, si->refcount);
   1272     }
   1273     return si->refcount;
   1274 }
   1275 
   1276 /* TODO: don't use unsigned for addrs below. It works, but is not
   1277  * ideal. They should probably be either uint32_t, Elf32_Addr, or unsigned
   1278  * long.
   1279  */
   1280 static int reloc_library(soinfo *si, Elf32_Rel *rel, unsigned count)
   1281 {
   1282     Elf32_Sym *symtab = si->symtab;
   1283     const char *strtab = si->strtab;
   1284     Elf32_Sym *s;
   1285     unsigned base;
   1286     Elf32_Rel *start = rel;
   1287     unsigned idx;
   1288 
   1289     for (idx = 0; idx < count; ++idx) {
   1290         unsigned type = ELF32_R_TYPE(rel->r_info);
   1291         unsigned sym = ELF32_R_SYM(rel->r_info);
   1292         unsigned reloc = (unsigned)(rel->r_offset + si->base);
   1293         unsigned sym_addr = 0;
   1294         char *sym_name = NULL;
   1295 
   1296         DEBUG("%5d Processing '%s' relocation at index %d\n", pid,
   1297               si->name, idx);
   1298         if(sym != 0) {
   1299             sym_name = (char *)(strtab + symtab[sym].st_name);
   1300             s = _do_lookup(si, sym_name, &base);
   1301             if(s == NULL) {
   1302                 /* We only allow an undefined symbol if this is a weak
   1303                    reference..   */
   1304                 s = &symtab[sym];
   1305                 if (ELF32_ST_BIND(s->st_info) != STB_WEAK) {
   1306                     DL_ERR("%5d cannot locate '%s'...\n", pid, sym_name);
   1307                     return -1;
   1308                 }
   1309 
   1310                 /* IHI0044C AAELF 4.5.1.1:
   1311 
   1312                    Libraries are not searched to resolve weak references.
   1313                    It is not an error for a weak reference to remain
   1314                    unsatisfied.
   1315 
   1316                    During linking, the value of an undefined weak reference is:
   1317                    - Zero if the relocation type is absolute
   1318                    - The address of the place if the relocation is pc-relative
   1319                    - The address of nominial base address if the relocation
   1320                      type is base-relative.
   1321                   */
   1322 
   1323                 switch (type) {
   1324 #if defined(ANDROID_ARM_LINKER)
   1325                 case R_ARM_JUMP_SLOT:
   1326                 case R_ARM_GLOB_DAT:
   1327                 case R_ARM_ABS32:
   1328                 case R_ARM_RELATIVE:    /* Don't care. */
   1329                 case R_ARM_NONE:        /* Don't care. */
   1330 #elif defined(ANDROID_X86_LINKER)
   1331                 case R_386_JUMP_SLOT:
   1332                 case R_386_GLOB_DAT:
   1333                 case R_386_32:
   1334                 case R_386_RELATIVE:    /* Dont' care. */
   1335 #endif /* ANDROID_*_LINKER */
   1336                     /* sym_addr was initialized to be zero above or relocation
   1337                        code below does not care about value of sym_addr.
   1338                        No need to do anything.  */
   1339                     break;
   1340 
   1341 #if defined(ANDROID_X86_LINKER)
   1342                 case R_386_PC32:
   1343                     sym_addr = reloc;
   1344                     break;
   1345 #endif /* ANDROID_X86_LINKER */
   1346 
   1347 #if defined(ANDROID_ARM_LINKER)
   1348                 case R_ARM_COPY:
   1349                     /* Fall through.  Can't really copy if weak symbol is
   1350                        not found in run-time.  */
   1351 #endif /* ANDROID_ARM_LINKER */
   1352                 default:
   1353                     DL_ERR("%5d unknown weak reloc type %d @ %p (%d)\n",
   1354                                  pid, type, rel, (int) (rel - start));
   1355                     return -1;
   1356                 }
   1357             } else {
   1358                 /* We got a definition.  */
   1359 #if 0
   1360             if((base == 0) && (si->base != 0)){
   1361                     /* linking from libraries to main image is bad */
   1362                 DL_ERR("%5d cannot locate '%s'...",
   1363                        pid, strtab + symtab[sym].st_name);
   1364                 return -1;
   1365             }
   1366 #endif
   1367                 sym_addr = (unsigned)(s->st_value + base);
   1368 	    }
   1369             COUNT_RELOC(RELOC_SYMBOL);
   1370         } else {
   1371             s = NULL;
   1372         }
   1373 
   1374 /* TODO: This is ugly. Split up the relocations by arch into
   1375  * different files.
   1376  */
   1377         switch(type){
   1378 #if defined(ANDROID_ARM_LINKER)
   1379         case R_ARM_JUMP_SLOT:
   1380             COUNT_RELOC(RELOC_ABSOLUTE);
   1381             MARK(rel->r_offset);
   1382             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1383                        reloc, sym_addr, sym_name);
   1384             *((unsigned*)reloc) = sym_addr;
   1385             break;
   1386         case R_ARM_GLOB_DAT:
   1387             COUNT_RELOC(RELOC_ABSOLUTE);
   1388             MARK(rel->r_offset);
   1389             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1390                        reloc, sym_addr, sym_name);
   1391             *((unsigned*)reloc) = sym_addr;
   1392             break;
   1393         case R_ARM_ABS32:
   1394             COUNT_RELOC(RELOC_ABSOLUTE);
   1395             MARK(rel->r_offset);
   1396             TRACE_TYPE(RELO, "%5d RELO ABS %08x <- %08x %s\n", pid,
   1397                        reloc, sym_addr, sym_name);
   1398             *((unsigned*)reloc) += sym_addr;
   1399             break;
   1400         case R_ARM_REL32:
   1401             COUNT_RELOC(RELOC_RELATIVE);
   1402             MARK(rel->r_offset);
   1403             TRACE_TYPE(RELO, "%5d RELO REL32 %08x <- %08x - %08x %s\n", pid,
   1404                        reloc, sym_addr, rel->r_offset, sym_name);
   1405             *((unsigned*)reloc) += sym_addr - rel->r_offset;
   1406             break;
   1407 #elif defined(ANDROID_X86_LINKER)
   1408         case R_386_JUMP_SLOT:
   1409             COUNT_RELOC(RELOC_ABSOLUTE);
   1410             MARK(rel->r_offset);
   1411             TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
   1412                        reloc, sym_addr, sym_name);
   1413             *((unsigned*)reloc) = sym_addr;
   1414             break;
   1415         case R_386_GLOB_DAT:
   1416             COUNT_RELOC(RELOC_ABSOLUTE);
   1417             MARK(rel->r_offset);
   1418             TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
   1419                        reloc, sym_addr, sym_name);
   1420             *((unsigned*)reloc) = sym_addr;
   1421             break;
   1422 #endif /* ANDROID_*_LINKER */
   1423 
   1424 #if defined(ANDROID_ARM_LINKER)
   1425         case R_ARM_RELATIVE:
   1426 #elif defined(ANDROID_X86_LINKER)
   1427         case R_386_RELATIVE:
   1428 #endif /* ANDROID_*_LINKER */
   1429             COUNT_RELOC(RELOC_RELATIVE);
   1430             MARK(rel->r_offset);
   1431             if(sym){
   1432                 DL_ERR("%5d odd RELATIVE form...", pid);
   1433                 return -1;
   1434             }
   1435             TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", pid,
   1436                        reloc, si->base);
   1437             *((unsigned*)reloc) += si->base;
   1438             break;
   1439 
   1440 #if defined(ANDROID_X86_LINKER)
   1441         case R_386_32:
   1442             COUNT_RELOC(RELOC_RELATIVE);
   1443             MARK(rel->r_offset);
   1444 
   1445             TRACE_TYPE(RELO, "%5d RELO R_386_32 %08x <- +%08x %s\n", pid,
   1446                        reloc, sym_addr, sym_name);
   1447             *((unsigned *)reloc) += (unsigned)sym_addr;
   1448             break;
   1449 
   1450         case R_386_PC32:
   1451             COUNT_RELOC(RELOC_RELATIVE);
   1452             MARK(rel->r_offset);
   1453             TRACE_TYPE(RELO, "%5d RELO R_386_PC32 %08x <- "
   1454                        "+%08x (%08x - %08x) %s\n", pid, reloc,
   1455                        (sym_addr - reloc), sym_addr, reloc, sym_name);
   1456             *((unsigned *)reloc) += (unsigned)(sym_addr - reloc);
   1457             break;
   1458 #endif /* ANDROID_X86_LINKER */
   1459 
   1460 #ifdef ANDROID_ARM_LINKER
   1461         case R_ARM_COPY:
   1462             COUNT_RELOC(RELOC_COPY);
   1463             MARK(rel->r_offset);
   1464             TRACE_TYPE(RELO, "%5d RELO %08x <- %d @ %08x %s\n", pid,
   1465                        reloc, s->st_size, sym_addr, sym_name);
   1466             memcpy((void*)reloc, (void*)sym_addr, s->st_size);
   1467             break;
   1468         case R_ARM_NONE:
   1469             break;
   1470 #endif /* ANDROID_ARM_LINKER */
   1471 
   1472         default:
   1473             DL_ERR("%5d unknown reloc type %d @ %p (%d)",
   1474                   pid, type, rel, (int) (rel - start));
   1475             return -1;
   1476         }
   1477         rel++;
   1478     }
   1479     return 0;
   1480 }
   1481 
   1482 /* Please read the "Initialization and Termination functions" functions.
   1483  * of the linker design note in bionic/linker/README.TXT to understand
   1484  * what the following code is doing.
   1485  *
   1486  * The important things to remember are:
   1487  *
   1488  *   DT_PREINIT_ARRAY must be called first for executables, and should
   1489  *   not appear in shared libraries.
   1490  *
   1491  *   DT_INIT should be called before DT_INIT_ARRAY if both are present
   1492  *
   1493  *   DT_FINI should be called after DT_FINI_ARRAY if both are present
   1494  *
   1495  *   DT_FINI_ARRAY must be parsed in reverse order.
   1496  */
   1497 
   1498 static void call_array(unsigned *ctor, int count, int reverse)
   1499 {
   1500     int n, inc = 1;
   1501 
   1502     if (reverse) {
   1503         ctor += (count-1);
   1504         inc   = -1;
   1505     }
   1506 
   1507     for(n = count; n > 0; n--) {
   1508         TRACE("[ %5d Looking at %s *0x%08x == 0x%08x ]\n", pid,
   1509               reverse ? "dtor" : "ctor",
   1510               (unsigned)ctor, (unsigned)*ctor);
   1511         void (*func)() = (void (*)()) *ctor;
   1512         ctor += inc;
   1513         if(((int) func == 0) || ((int) func == -1)) continue;
   1514         TRACE("[ %5d Calling func @ 0x%08x ]\n", pid, (unsigned)func);
   1515         func();
   1516     }
   1517 }
   1518 
   1519 void call_constructors_recursive(soinfo *si)
   1520 {
   1521     if (si->constructors_called)
   1522         return;
   1523 
   1524     // Set this before actually calling the constructors, otherwise it doesn't
   1525     // protect against recursive constructor calls. One simple example of
   1526     // constructor recursion is the libc debug malloc, which is implemented in
   1527     // libc_malloc_debug_leak.so:
   1528     // 1. The program depends on libc, so libc's constructor is called here.
   1529     // 2. The libc constructor calls dlopen() to load libc_malloc_debug_leak.so.
   1530     // 3. dlopen() calls call_constructors_recursive() with the newly created
   1531     //    soinfo for libc_malloc_debug_leak.so.
   1532     // 4. The debug so depends on libc, so call_constructors_recursive() is
   1533     //    called again with the libc soinfo. If it doesn't trigger the early-
   1534     //    out above, the libc constructor will be called again (recursively!).
   1535     si->constructors_called = 1;
   1536 
   1537     if (si->flags & FLAG_EXE) {
   1538         TRACE("[ %5d Calling preinit_array @ 0x%08x [%d] for '%s' ]\n",
   1539               pid, (unsigned)si->preinit_array, si->preinit_array_count,
   1540               si->name);
   1541         call_array(si->preinit_array, si->preinit_array_count, 0);
   1542         TRACE("[ %5d Done calling preinit_array for '%s' ]\n", pid, si->name);
   1543     } else {
   1544         if (si->preinit_array) {
   1545             DL_ERR("%5d Shared library '%s' has a preinit_array table @ 0x%08x."
   1546                    " This is INVALID.", pid, si->name,
   1547                    (unsigned)si->preinit_array);
   1548         }
   1549     }
   1550 
   1551     if (si->dynamic) {
   1552         unsigned *d;
   1553         for(d = si->dynamic; *d; d += 2) {
   1554             if(d[0] == DT_NEEDED){
   1555                 soinfo* lsi = (soinfo *)d[1];
   1556                 if (!validate_soinfo(lsi)) {
   1557                     DL_ERR("%5d bad DT_NEEDED pointer in %s",
   1558                            pid, si->name);
   1559                 } else {
   1560                     call_constructors_recursive(lsi);
   1561                 }
   1562             }
   1563         }
   1564     }
   1565 
   1566     if (si->init_func) {
   1567         TRACE("[ %5d Calling init_func @ 0x%08x for '%s' ]\n", pid,
   1568               (unsigned)si->init_func, si->name);
   1569         si->init_func();
   1570         TRACE("[ %5d Done calling init_func for '%s' ]\n", pid, si->name);
   1571     }
   1572 
   1573     if (si->init_array) {
   1574         TRACE("[ %5d Calling init_array @ 0x%08x [%d] for '%s' ]\n", pid,
   1575               (unsigned)si->init_array, si->init_array_count, si->name);
   1576         call_array(si->init_array, si->init_array_count, 0);
   1577         TRACE("[ %5d Done calling init_array for '%s' ]\n", pid, si->name);
   1578     }
   1579 
   1580 }
   1581 
   1582 static void call_destructors(soinfo *si)
   1583 {
   1584     if (si->fini_array) {
   1585         TRACE("[ %5d Calling fini_array @ 0x%08x [%d] for '%s' ]\n", pid,
   1586               (unsigned)si->fini_array, si->fini_array_count, si->name);
   1587         call_array(si->fini_array, si->fini_array_count, 1);
   1588         TRACE("[ %5d Done calling fini_array for '%s' ]\n", pid, si->name);
   1589     }
   1590 
   1591     if (si->fini_func) {
   1592         TRACE("[ %5d Calling fini_func @ 0x%08x for '%s' ]\n", pid,
   1593               (unsigned)si->fini_func, si->name);
   1594         si->fini_func();
   1595         TRACE("[ %5d Done calling fini_func for '%s' ]\n", pid, si->name);
   1596     }
   1597 }
   1598 
   1599 /* Force any of the closed stdin, stdout and stderr to be associated with
   1600    /dev/null. */
   1601 static int nullify_closed_stdio (void)
   1602 {
   1603     int dev_null, i, status;
   1604     int return_value = 0;
   1605 
   1606     dev_null = open("/dev/null", O_RDWR);
   1607     if (dev_null < 0) {
   1608         DL_ERR("Cannot open /dev/null.");
   1609         return -1;
   1610     }
   1611     TRACE("[ %5d Opened /dev/null file-descriptor=%d]\n", pid, dev_null);
   1612 
   1613     /* If any of the stdio file descriptors is valid and not associated
   1614        with /dev/null, dup /dev/null to it.  */
   1615     for (i = 0; i < 3; i++) {
   1616         /* If it is /dev/null already, we are done. */
   1617         if (i == dev_null)
   1618             continue;
   1619 
   1620         TRACE("[ %5d Nullifying stdio file descriptor %d]\n", pid, i);
   1621         /* The man page of fcntl does not say that fcntl(..,F_GETFL)
   1622            can be interrupted but we do this just to be safe. */
   1623         do {
   1624           status = fcntl(i, F_GETFL);
   1625         } while (status < 0 && errno == EINTR);
   1626 
   1627         /* If file is openned, we are good. */
   1628         if (status >= 0)
   1629           continue;
   1630 
   1631         /* The only error we allow is that the file descriptor does not
   1632            exist, in which case we dup /dev/null to it. */
   1633         if (errno != EBADF) {
   1634             DL_ERR("nullify_stdio: unhandled error %s", strerror(errno));
   1635             return_value = -1;
   1636             continue;
   1637         }
   1638 
   1639         /* Try dupping /dev/null to this stdio file descriptor and
   1640            repeat if there is a signal.  Note that any errors in closing
   1641            the stdio descriptor are lost.  */
   1642         do {
   1643             status = dup2(dev_null, i);
   1644         } while (status < 0 && errno == EINTR);
   1645 
   1646         if (status < 0) {
   1647             DL_ERR("nullify_stdio: dup2 error %s", strerror(errno));
   1648             return_value = -1;
   1649             continue;
   1650         }
   1651     }
   1652 
   1653     /* If /dev/null is not one of the stdio file descriptors, close it. */
   1654     if (dev_null > 2) {
   1655         TRACE("[ %5d Closing /dev/null file-descriptor=%d]\n", pid, dev_null);
   1656         do {
   1657             status = close(dev_null);
   1658         } while (status < 0 && errno == EINTR);
   1659 
   1660         if (status < 0) {
   1661             DL_ERR("nullify_stdio: close error %s", strerror(errno));
   1662             return_value = -1;
   1663         }
   1664     }
   1665 
   1666     return return_value;
   1667 }
   1668 
   1669 static int link_image(soinfo *si, unsigned wr_offset)
   1670 {
   1671     unsigned *d;
   1672     Elf32_Phdr *phdr = si->phdr;
   1673     int phnum = si->phnum;
   1674 
   1675     INFO("[ %5d linking %s ]\n", pid, si->name);
   1676     DEBUG("%5d si->base = 0x%08x si->flags = 0x%08x\n", pid,
   1677           si->base, si->flags);
   1678 
   1679     if (si->flags & (FLAG_EXE | FLAG_LINKER)) {
   1680         /* Locate the needed program segments (DYNAMIC/ARM_EXIDX) for
   1681          * linkage info if this is the executable or the linker itself.
   1682          * If this was a dynamic lib, that would have been done at load time.
   1683          *
   1684          * TODO: It's unfortunate that small pieces of this are
   1685          * repeated from the load_library routine. Refactor this just
   1686          * slightly to reuse these bits.
   1687          */
   1688         si->size = 0;
   1689         for(; phnum > 0; --phnum, ++phdr) {
   1690 #ifdef ANDROID_ARM_LINKER
   1691             if(phdr->p_type == PT_ARM_EXIDX) {
   1692                 /* exidx entries (used for stack unwinding) are 8 bytes each.
   1693                  */
   1694                 si->ARM_exidx = (unsigned *)phdr->p_vaddr;
   1695                 si->ARM_exidx_count = phdr->p_memsz / 8;
   1696             }
   1697 #endif
   1698             if (phdr->p_type == PT_LOAD) {
   1699                 /* For the executable, we use the si->size field only in
   1700                    dl_unwind_find_exidx(), so the meaning of si->size
   1701                    is not the size of the executable; it is the distance
   1702                    between the load location of the executable and the last
   1703                    address of the loadable part of the executable.
   1704                    We use the range [si->base, si->base + si->size) to
   1705                    determine whether a PC value falls within the executable
   1706                    section. Of course, if a value is between si->base and
   1707                    (si->base + phdr->p_vaddr), it's not in the executable
   1708                    section, but a) we shouldn't be asking for such a value
   1709                    anyway, and b) if we have to provide an EXIDX for such a
   1710                    value, then the executable's EXIDX is probably the better
   1711                    choice.
   1712                 */
   1713                 DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
   1714                 if (phdr->p_vaddr + phdr->p_memsz > si->size)
   1715                     si->size = phdr->p_vaddr + phdr->p_memsz;
   1716                 /* try to remember what range of addresses should be write
   1717                  * protected */
   1718                 if (!(phdr->p_flags & PF_W)) {
   1719                     unsigned _end;
   1720 
   1721                     if (si->base + phdr->p_vaddr < si->wrprotect_start)
   1722                         si->wrprotect_start = si->base + phdr->p_vaddr;
   1723                     _end = (((si->base + phdr->p_vaddr + phdr->p_memsz + PAGE_SIZE - 1) &
   1724                              (~PAGE_MASK)));
   1725                     if (_end > si->wrprotect_end)
   1726                         si->wrprotect_end = _end;
   1727                     /* Make the section writable just in case we'll have to
   1728                      * write to it during relocation (i.e. text segment).
   1729                      * However, we will remember what range of addresses
   1730                      * should be write protected.
   1731                      */
   1732                     mprotect((void *) (si->base + phdr->p_vaddr),
   1733                              phdr->p_memsz,
   1734                              PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
   1735                 }
   1736             } else if (phdr->p_type == PT_DYNAMIC) {
   1737                 if (si->dynamic != (unsigned *)-1) {
   1738                     DL_ERR("%5d multiple PT_DYNAMIC segments found in '%s'. "
   1739                           "Segment at 0x%08x, previously one found at 0x%08x",
   1740                           pid, si->name, si->base + phdr->p_vaddr,
   1741                           (unsigned)si->dynamic);
   1742                     goto fail;
   1743                 }
   1744                 DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
   1745                 si->dynamic = (unsigned *) (si->base + phdr->p_vaddr);
   1746             } else if (phdr->p_type == PT_GNU_RELRO) {
   1747                 if ((phdr->p_vaddr >= si->size)
   1748                         || ((phdr->p_vaddr + phdr->p_memsz) > si->size)
   1749                         || ((si->base + phdr->p_vaddr + phdr->p_memsz) < si->base)) {
   1750                     DL_ERR("%d invalid GNU_RELRO in '%s' "
   1751                            "p_vaddr=0x%08x p_memsz=0x%08x", pid, si->name,
   1752                            phdr->p_vaddr, phdr->p_memsz);
   1753                     goto fail;
   1754                 }
   1755                 si->gnu_relro_start = (Elf32_Addr) (si->base + phdr->p_vaddr);
   1756                 si->gnu_relro_len = (unsigned) phdr->p_memsz;
   1757             }
   1758         }
   1759     }
   1760 
   1761     if (si->dynamic == (unsigned *)-1) {
   1762         DL_ERR("%5d missing PT_DYNAMIC?!", pid);
   1763         goto fail;
   1764     }
   1765 
   1766     DEBUG("%5d dynamic = %p\n", pid, si->dynamic);
   1767 
   1768     /* extract useful information from dynamic section */
   1769     for(d = si->dynamic; *d; d++){
   1770         DEBUG("%5d d = %p, d[0] = 0x%08x d[1] = 0x%08x\n", pid, d, d[0], d[1]);
   1771         switch(*d++){
   1772         case DT_HASH:
   1773             si->nbucket = ((unsigned *) (si->base + *d))[0];
   1774             si->nchain = ((unsigned *) (si->base + *d))[1];
   1775             si->bucket = (unsigned *) (si->base + *d + 8);
   1776             si->chain = (unsigned *) (si->base + *d + 8 + si->nbucket * 4);
   1777             break;
   1778         case DT_STRTAB:
   1779             si->strtab = (const char *) (si->base + *d);
   1780             break;
   1781         case DT_SYMTAB:
   1782             si->symtab = (Elf32_Sym *) (si->base + *d);
   1783             break;
   1784         case DT_PLTREL:
   1785             if(*d != DT_REL) {
   1786                 DL_ERR("DT_RELA not supported");
   1787                 goto fail;
   1788             }
   1789             break;
   1790         case DT_JMPREL:
   1791             si->plt_rel = (Elf32_Rel*) (si->base + *d);
   1792             break;
   1793         case DT_PLTRELSZ:
   1794             si->plt_rel_count = *d / 8;
   1795             break;
   1796         case DT_REL:
   1797             si->rel = (Elf32_Rel*) (si->base + *d);
   1798             break;
   1799         case DT_RELSZ:
   1800             si->rel_count = *d / 8;
   1801             break;
   1802         case DT_PLTGOT:
   1803             /* Save this in case we decide to do lazy binding. We don't yet. */
   1804             si->plt_got = (unsigned *)(si->base + *d);
   1805             break;
   1806         case DT_DEBUG:
   1807             // Set the DT_DEBUG entry to the addres of _r_debug for GDB
   1808             *d = (int) &_r_debug;
   1809             break;
   1810          case DT_RELA:
   1811             DL_ERR("%5d DT_RELA not supported", pid);
   1812             goto fail;
   1813         case DT_INIT:
   1814             si->init_func = (void (*)(void))(si->base + *d);
   1815             DEBUG("%5d %s constructors (init func) found at %p\n",
   1816                   pid, si->name, si->init_func);
   1817             break;
   1818         case DT_FINI:
   1819             si->fini_func = (void (*)(void))(si->base + *d);
   1820             DEBUG("%5d %s destructors (fini func) found at %p\n",
   1821                   pid, si->name, si->fini_func);
   1822             break;
   1823         case DT_INIT_ARRAY:
   1824             si->init_array = (unsigned *)(si->base + *d);
   1825             DEBUG("%5d %s constructors (init_array) found at %p\n",
   1826                   pid, si->name, si->init_array);
   1827             break;
   1828         case DT_INIT_ARRAYSZ:
   1829             si->init_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1830             break;
   1831         case DT_FINI_ARRAY:
   1832             si->fini_array = (unsigned *)(si->base + *d);
   1833             DEBUG("%5d %s destructors (fini_array) found at %p\n",
   1834                   pid, si->name, si->fini_array);
   1835             break;
   1836         case DT_FINI_ARRAYSZ:
   1837             si->fini_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1838             break;
   1839         case DT_PREINIT_ARRAY:
   1840             si->preinit_array = (unsigned *)(si->base + *d);
   1841             DEBUG("%5d %s constructors (preinit_array) found at %p\n",
   1842                   pid, si->name, si->preinit_array);
   1843             break;
   1844         case DT_PREINIT_ARRAYSZ:
   1845             si->preinit_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
   1846             break;
   1847         case DT_TEXTREL:
   1848             /* TODO: make use of this. */
   1849             /* this means that we might have to write into where the text
   1850              * segment was loaded during relocation... Do something with
   1851              * it.
   1852              */
   1853             DEBUG("%5d Text segment should be writable during relocation.\n",
   1854                   pid);
   1855             break;
   1856         }
   1857     }
   1858 
   1859     DEBUG("%5d si->base = 0x%08x, si->strtab = %p, si->symtab = %p\n",
   1860            pid, si->base, si->strtab, si->symtab);
   1861 
   1862     if((si->strtab == 0) || (si->symtab == 0)) {
   1863         DL_ERR("%5d missing essential tables", pid);
   1864         goto fail;
   1865     }
   1866 
   1867     /* if this is the main executable, then load all of the preloads now */
   1868     if(si->flags & FLAG_EXE) {
   1869         int i;
   1870         memset(preloads, 0, sizeof(preloads));
   1871         for(i = 0; ldpreload_names[i] != NULL; i++) {
   1872             soinfo *lsi = find_library(ldpreload_names[i]);
   1873             if(lsi == 0) {
   1874                 strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
   1875                 DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
   1876                        pid, ldpreload_names[i], si->name, tmp_err_buf);
   1877                 goto fail;
   1878             }
   1879             lsi->refcount++;
   1880             preloads[i] = lsi;
   1881         }
   1882     }
   1883 
   1884     for(d = si->dynamic; *d; d += 2) {
   1885         if(d[0] == DT_NEEDED){
   1886             DEBUG("%5d %s needs %s\n", pid, si->name, si->strtab + d[1]);
   1887             soinfo *lsi = find_library(si->strtab + d[1]);
   1888             if(lsi == 0) {
   1889                 strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
   1890                 DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
   1891                        pid, si->strtab + d[1], si->name, tmp_err_buf);
   1892                 goto fail;
   1893             }
   1894             /* Save the soinfo of the loaded DT_NEEDED library in the payload
   1895                of the DT_NEEDED entry itself, so that we can retrieve the
   1896                soinfo directly later from the dynamic segment.  This is a hack,
   1897                but it allows us to map from DT_NEEDED to soinfo efficiently
   1898                later on when we resolve relocations, trying to look up a symbol
   1899                with dlsym().
   1900             */
   1901             d[1] = (unsigned)lsi;
   1902             lsi->refcount++;
   1903         }
   1904     }
   1905 
   1906     if(si->plt_rel) {
   1907         DEBUG("[ %5d relocating %s plt ]\n", pid, si->name );
   1908         if(reloc_library(si, si->plt_rel, si->plt_rel_count))
   1909             goto fail;
   1910     }
   1911     if(si->rel) {
   1912         DEBUG("[ %5d relocating %s ]\n", pid, si->name );
   1913         if(reloc_library(si, si->rel, si->rel_count))
   1914             goto fail;
   1915     }
   1916 
   1917     si->flags |= FLAG_LINKED;
   1918     DEBUG("[ %5d finished linking %s ]\n", pid, si->name);
   1919 
   1920 #if 0
   1921     /* This is the way that the old dynamic linker did protection of
   1922      * non-writable areas. It would scan section headers and find where
   1923      * .text ended (rather where .data/.bss began) and assume that this is
   1924      * the upper range of the non-writable area. This is too coarse,
   1925      * and is kept here for reference until we fully move away from single
   1926      * segment elf objects. See the code in get_wr_offset (also #if'd 0)
   1927      * that made this possible.
   1928      */
   1929     if(wr_offset < 0xffffffff){
   1930         mprotect((void*) si->base, wr_offset, PROT_READ | PROT_EXEC);
   1931     }
   1932 #else
   1933     /* TODO: Verify that this does the right thing in all cases, as it
   1934      * presently probably does not. It is possible that an ELF image will
   1935      * come with multiple read-only segments. What we ought to do is scan
   1936      * the program headers again and mprotect all the read-only segments.
   1937      * To prevent re-scanning the program header, we would have to build a
   1938      * list of loadable segments in si, and then scan that instead. */
   1939     if (si->wrprotect_start != 0xffffffff && si->wrprotect_end != 0) {
   1940         mprotect((void *)si->wrprotect_start,
   1941                  si->wrprotect_end - si->wrprotect_start,
   1942                  PROT_READ | PROT_EXEC);
   1943     }
   1944 #endif
   1945 
   1946     if (si->gnu_relro_start != 0 && si->gnu_relro_len != 0) {
   1947         Elf32_Addr start = (si->gnu_relro_start & ~PAGE_MASK);
   1948         unsigned len = (si->gnu_relro_start - start) + si->gnu_relro_len;
   1949         if (mprotect((void *) start, len, PROT_READ) < 0) {
   1950             DL_ERR("%5d GNU_RELRO mprotect of library '%s' failed: %d (%s)\n",
   1951                    pid, si->name, errno, strerror(errno));
   1952             goto fail;
   1953         }
   1954     }
   1955 
   1956     /* If this is a SET?ID program, dup /dev/null to opened stdin,
   1957        stdout and stderr to close a security hole described in:
   1958 
   1959     ftp://ftp.freebsd.org/pub/FreeBSD/CERT/advisories/FreeBSD-SA-02:23.stdio.asc
   1960 
   1961      */
   1962     if (program_is_setuid)
   1963         nullify_closed_stdio ();
   1964     notify_gdb_of_load(si);
   1965     return 0;
   1966 
   1967 fail:
   1968     ERROR("failed to link %s\n", si->name);
   1969     si->flags |= FLAG_ERROR;
   1970     return -1;
   1971 }
   1972 
   1973 static void parse_library_path(const char *path, char *delim)
   1974 {
   1975     size_t len;
   1976     char *ldpaths_bufp = ldpaths_buf;
   1977     int i = 0;
   1978 
   1979     len = strlcpy(ldpaths_buf, path, sizeof(ldpaths_buf));
   1980 
   1981     while (i < LDPATH_MAX && (ldpaths[i] = strsep(&ldpaths_bufp, delim))) {
   1982         if (*ldpaths[i] != '\0')
   1983             ++i;
   1984     }
   1985 
   1986     /* Forget the last path if we had to truncate; this occurs if the 2nd to
   1987      * last char isn't '\0' (i.e. not originally a delim). */
   1988     if (i > 0 && len >= sizeof(ldpaths_buf) &&
   1989             ldpaths_buf[sizeof(ldpaths_buf) - 2] != '\0') {
   1990         ldpaths[i - 1] = NULL;
   1991     } else {
   1992         ldpaths[i] = NULL;
   1993     }
   1994 }
   1995 
   1996 static void parse_preloads(const char *path, char *delim)
   1997 {
   1998     size_t len;
   1999     char *ldpreloads_bufp = ldpreloads_buf;
   2000     int i = 0;
   2001 
   2002     len = strlcpy(ldpreloads_buf, path, sizeof(ldpreloads_buf));
   2003 
   2004     while (i < LDPRELOAD_MAX && (ldpreload_names[i] = strsep(&ldpreloads_bufp, delim))) {
   2005         if (*ldpreload_names[i] != '\0') {
   2006             ++i;
   2007         }
   2008     }
   2009 
   2010     /* Forget the last path if we had to truncate; this occurs if the 2nd to
   2011      * last char isn't '\0' (i.e. not originally a delim). */
   2012     if (i > 0 && len >= sizeof(ldpreloads_buf) &&
   2013             ldpreloads_buf[sizeof(ldpreloads_buf) - 2] != '\0') {
   2014         ldpreload_names[i - 1] = NULL;
   2015     } else {
   2016         ldpreload_names[i] = NULL;
   2017     }
   2018 }
   2019 
   2020 /*
   2021  * This code is called after the linker has linked itself and
   2022  * fixed it's own GOT. It is safe to make references to externs
   2023  * and other non-local data at this point.
   2024  */
   2025 static unsigned __linker_init_post_relocation(unsigned **elfdata)
   2026 {
   2027     static soinfo linker_soinfo;
   2028 
   2029     int argc = (int) *elfdata;
   2030     char **argv = (char**) (elfdata + 1);
   2031     unsigned *vecs = (unsigned*) (argv + argc + 1);
   2032     unsigned *v;
   2033     soinfo *si;
   2034     struct link_map * map;
   2035     const char *ldpath_env = NULL;
   2036     const char *ldpreload_env = NULL;
   2037 
   2038     /* NOTE: we store the elfdata pointer on a special location
   2039      *       of the temporary TLS area in order to pass it to
   2040      *       the C Library's runtime initializer.
   2041      *
   2042      *       The initializer must clear the slot and reset the TLS
   2043      *       to point to a different location to ensure that no other
   2044      *       shared library constructor can access it.
   2045      */
   2046     __libc_init_tls(elfdata);
   2047 
   2048     pid = getpid();
   2049 
   2050 #if TIMING
   2051     struct timeval t0, t1;
   2052     gettimeofday(&t0, 0);
   2053 #endif
   2054 
   2055     /* Initialize environment functions, and get to the ELF aux vectors table */
   2056     vecs = linker_env_init(vecs);
   2057 
   2058     /* Check auxv for AT_SECURE first to see if program is setuid, setgid,
   2059        has file caps, or caused a SELinux/AppArmor domain transition. */
   2060     for (v = vecs; v[0]; v += 2) {
   2061         if (v[0] == AT_SECURE) {
   2062             /* kernel told us whether to enable secure mode */
   2063             program_is_setuid = v[1];
   2064             goto sanitize;
   2065         }
   2066     }
   2067 
   2068     /* Kernel did not provide AT_SECURE - fall back on legacy test. */
   2069     program_is_setuid = (getuid() != geteuid()) || (getgid() != getegid());
   2070 
   2071 sanitize:
   2072     /* Sanitize environment if we're loading a setuid program */
   2073     if (program_is_setuid)
   2074         linker_env_secure();
   2075 
   2076     debugger_init();
   2077 
   2078     /* Get a few environment variables */
   2079     {
   2080 #if LINKER_DEBUG
   2081         const char* env;
   2082         env = linker_env_get("DEBUG"); /* XXX: TODO: Change to LD_DEBUG */
   2083         if (env)
   2084             debug_verbosity = atoi(env);
   2085 #endif
   2086 
   2087         /* Normally, these are cleaned by linker_env_secure, but the test
   2088          * against program_is_setuid doesn't cost us anything */
   2089         if (!program_is_setuid) {
   2090             ldpath_env = linker_env_get("LD_LIBRARY_PATH");
   2091             ldpreload_env = linker_env_get("LD_PRELOAD");
   2092         }
   2093     }
   2094 
   2095     INFO("[ android linker & debugger ]\n");
   2096     DEBUG("%5d elfdata @ 0x%08x\n", pid, (unsigned)elfdata);
   2097 
   2098     si = alloc_info(argv[0]);
   2099     if(si == 0) {
   2100         exit(-1);
   2101     }
   2102 
   2103         /* bootstrap the link map, the main exe always needs to be first */
   2104     si->flags |= FLAG_EXE;
   2105     map = &(si->linkmap);
   2106 
   2107     map->l_addr = 0;
   2108     map->l_name = argv[0];
   2109     map->l_prev = NULL;
   2110     map->l_next = NULL;
   2111 
   2112     _r_debug.r_map = map;
   2113     r_debug_tail = map;
   2114 
   2115         /* gdb expects the linker to be in the debug shared object list,
   2116          * and we need to make sure that the reported load address is zero.
   2117          * Without this, gdb gets the wrong idea of where rtld_db_dlactivity()
   2118          * is.  Don't use alloc_info(), because the linker shouldn't
   2119          * be on the soinfo list.
   2120          */
   2121     strlcpy((char*) linker_soinfo.name, "/system/bin/linker", sizeof linker_soinfo.name);
   2122     linker_soinfo.flags = 0;
   2123     linker_soinfo.base = 0;     // This is the important part; must be zero.
   2124     insert_soinfo_into_debug_map(&linker_soinfo);
   2125 
   2126         /* extract information passed from the kernel */
   2127     while(vecs[0] != 0){
   2128         switch(vecs[0]){
   2129         case AT_PHDR:
   2130             si->phdr = (Elf32_Phdr*) vecs[1];
   2131             break;
   2132         case AT_PHNUM:
   2133             si->phnum = (int) vecs[1];
   2134             break;
   2135         case AT_ENTRY:
   2136             si->entry = vecs[1];
   2137             break;
   2138         }
   2139         vecs += 2;
   2140     }
   2141 
   2142     /* Compute the value of si->base. We can't rely on the fact that
   2143      * the first entry is the PHDR because this will not be true
   2144      * for certain executables (e.g. some in the NDK unit test suite)
   2145      */
   2146     int nn;
   2147     si->base = 0;
   2148     for ( nn = 0; nn < si->phnum; nn++ ) {
   2149         if (si->phdr[nn].p_type == PT_PHDR) {
   2150             si->base = (Elf32_Addr) si->phdr - si->phdr[nn].p_vaddr;
   2151             break;
   2152         }
   2153     }
   2154     si->dynamic = (unsigned *)-1;
   2155     si->wrprotect_start = 0xffffffff;
   2156     si->wrprotect_end = 0;
   2157     si->refcount = 1;
   2158     si->gnu_relro_start = 0;
   2159     si->gnu_relro_len = 0;
   2160 
   2161         /* Use LD_LIBRARY_PATH if we aren't setuid/setgid */
   2162     if (ldpath_env)
   2163         parse_library_path(ldpath_env, ":");
   2164 
   2165     if (ldpreload_env) {
   2166         parse_preloads(ldpreload_env, " :");
   2167     }
   2168 
   2169     if(link_image(si, 0)) {
   2170         char errmsg[] = "CANNOT LINK EXECUTABLE\n";
   2171         write(2, __linker_dl_err_buf, strlen(__linker_dl_err_buf));
   2172         write(2, errmsg, sizeof(errmsg));
   2173         exit(-1);
   2174     }
   2175 
   2176     call_constructors_recursive(si);
   2177 
   2178 #if ALLOW_SYMBOLS_FROM_MAIN
   2179     /* Set somain after we've loaded all the libraries in order to prevent
   2180      * linking of symbols back to the main image, which is not set up at that
   2181      * point yet.
   2182      */
   2183     somain = si;
   2184 #endif
   2185 
   2186 #if TIMING
   2187     gettimeofday(&t1,NULL);
   2188     PRINT("LINKER TIME: %s: %d microseconds\n", argv[0], (int) (
   2189                (((long long)t1.tv_sec * 1000000LL) + (long long)t1.tv_usec) -
   2190                (((long long)t0.tv_sec * 1000000LL) + (long long)t0.tv_usec)
   2191                ));
   2192 #endif
   2193 #if STATS
   2194     PRINT("RELO STATS: %s: %d abs, %d rel, %d copy, %d symbol\n", argv[0],
   2195            linker_stats.reloc[RELOC_ABSOLUTE],
   2196            linker_stats.reloc[RELOC_RELATIVE],
   2197            linker_stats.reloc[RELOC_COPY],
   2198            linker_stats.reloc[RELOC_SYMBOL]);
   2199 #endif
   2200 #if COUNT_PAGES
   2201     {
   2202         unsigned n;
   2203         unsigned i;
   2204         unsigned count = 0;
   2205         for(n = 0; n < 4096; n++){
   2206             if(bitmask[n]){
   2207                 unsigned x = bitmask[n];
   2208                 for(i = 0; i < 8; i++){
   2209                     if(x & 1) count++;
   2210                     x >>= 1;
   2211                 }
   2212             }
   2213         }
   2214         PRINT("PAGES MODIFIED: %s: %d (%dKB)\n", argv[0], count, count * 4);
   2215     }
   2216 #endif
   2217 
   2218 #if TIMING || STATS || COUNT_PAGES
   2219     fflush(stdout);
   2220 #endif
   2221 
   2222     TRACE("[ %5d Ready to execute '%s' @ 0x%08x ]\n", pid, si->name,
   2223           si->entry);
   2224     return si->entry;
   2225 }
   2226 
   2227 /*
   2228  * Find the value of AT_BASE passed to us by the kernel. This is the load
   2229  * location of the linker.
   2230  */
   2231 static unsigned find_linker_base(unsigned **elfdata) {
   2232     int argc = (int) *elfdata;
   2233     char **argv = (char**) (elfdata + 1);
   2234     unsigned *vecs = (unsigned*) (argv + argc + 1);
   2235     while (vecs[0] != 0) {
   2236         vecs++;
   2237     }
   2238 
   2239     /* The end of the environment block is marked by two NULL pointers */
   2240     vecs++;
   2241 
   2242     while(vecs[0]) {
   2243         if (vecs[0] == AT_BASE) {
   2244             return vecs[1];
   2245         }
   2246         vecs += 2;
   2247     }
   2248 
   2249     return 0; // should never happen
   2250 }
   2251 
   2252 /*
   2253  * This is the entry point for the linker, called from begin.S. This
   2254  * method is responsible for fixing the linker's own relocations, and
   2255  * then calling __linker_init_post_relocation().
   2256  *
   2257  * Because this method is called before the linker has fixed it's own
   2258  * relocations, any attempt to reference an extern variable, extern
   2259  * function, or other GOT reference will generate a segfault.
   2260  */
   2261 unsigned __linker_init(unsigned **elfdata) {
   2262     unsigned linker_addr = find_linker_base(elfdata);
   2263     Elf32_Ehdr *elf_hdr = (Elf32_Ehdr *) linker_addr;
   2264     Elf32_Phdr *phdr =
   2265         (Elf32_Phdr *)((unsigned char *) linker_addr + elf_hdr->e_phoff);
   2266 
   2267     soinfo linker_so;
   2268     memset(&linker_so, 0, sizeof(soinfo));
   2269 
   2270     linker_so.base = linker_addr;
   2271     linker_so.dynamic = (unsigned *) -1;
   2272     linker_so.phdr = phdr;
   2273     linker_so.phnum = elf_hdr->e_phnum;
   2274     linker_so.flags |= FLAG_LINKER;
   2275     linker_so.wrprotect_start = 0xffffffff;
   2276     linker_so.wrprotect_end = 0;
   2277     linker_so.gnu_relro_start = 0;
   2278     linker_so.gnu_relro_len = 0;
   2279 
   2280     if (link_image(&linker_so, 0)) {
   2281         // It would be nice to print an error message, but if the linker
   2282         // can't link itself, there's no guarantee that we'll be able to
   2283         // call write() (because it involves a GOT reference).
   2284         //
   2285         // This situation should never occur unless the linker itself
   2286         // is corrupt.
   2287         exit(-1);
   2288     }
   2289 
   2290     // We have successfully fixed our own relocations. It's safe to run
   2291     // the main part of the linker now.
   2292     return __linker_init_post_relocation(elfdata);
   2293 }
   2294