Home | History | Annotate | Download | only in libbacktrace
      1 /*
      2  * Copyright (C) 2013 The Android Open Source Project
      3  *
      4  * Licensed under the Apache License, Version 2.0 (the "License");
      5  * you may not use this file except in compliance with the License.
      6  * You may obtain a copy of the License at
      7  *
      8  *      http://www.apache.org/licenses/LICENSE-2.0
      9  *
     10  * Unless required by applicable law or agreed to in writing, software
     11  * distributed under the License is distributed on an "AS IS" BASIS,
     12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     13  * See the License for the specific language governing permissions and
     14  * limitations under the License.
     15  */
     16 
     17 #include <dirent.h>
     18 #include <errno.h>
     19 #include <inttypes.h>
     20 #include <pthread.h>
     21 #include <signal.h>
     22 #include <stdint.h>
     23 #include <stdio.h>
     24 #include <stdlib.h>
     25 #include <string.h>
     26 #include <sys/ptrace.h>
     27 #include <sys/types.h>
     28 #include <sys/wait.h>
     29 #include <time.h>
     30 #include <unistd.h>
     31 
     32 #include <backtrace/Backtrace.h>
     33 #include <backtrace/BacktraceMap.h>
     34 #include <UniquePtr.h>
     35 
     36 // For the THREAD_SIGNAL definition.
     37 #include "BacktraceThread.h"
     38 
     39 #include <cutils/atomic.h>
     40 #include <gtest/gtest.h>
     41 
     42 #include <algorithm>
     43 #include <vector>
     44 
     45 #include "thread_utils.h"
     46 
     47 // Number of microseconds per milliseconds.
     48 #define US_PER_MSEC             1000
     49 
     50 // Number of nanoseconds in a second.
     51 #define NS_PER_SEC              1000000000ULL
     52 
     53 // Number of simultaneous dumping operations to perform.
     54 #define NUM_THREADS  40
     55 
     56 // Number of simultaneous threads running in our forked process.
     57 #define NUM_PTRACE_THREADS 5
     58 
     59 struct thread_t {
     60   pid_t tid;
     61   int32_t state;
     62   pthread_t threadId;
     63 };
     64 
     65 struct dump_thread_t {
     66   thread_t thread;
     67   Backtrace* backtrace;
     68   int32_t* now;
     69   int32_t done;
     70 };
     71 
     72 extern "C" {
     73 // Prototypes for functions in the test library.
     74 int test_level_one(int, int, int, int, void (*)(void*), void*);
     75 
     76 int test_recursive_call(int, void (*)(void*), void*);
     77 }
     78 
     79 uint64_t NanoTime() {
     80   struct timespec t = { 0, 0 };
     81   clock_gettime(CLOCK_MONOTONIC, &t);
     82   return static_cast<uint64_t>(t.tv_sec * NS_PER_SEC + t.tv_nsec);
     83 }
     84 
     85 void DumpFrames(Backtrace* backtrace) {
     86   if (backtrace->NumFrames() == 0) {
     87     printf("    No frames to dump\n");
     88     return;
     89   }
     90 
     91   for (size_t i = 0; i < backtrace->NumFrames(); i++) {
     92     printf("    %s\n", backtrace->FormatFrameData(i).c_str());
     93   }
     94 }
     95 
     96 void WaitForStop(pid_t pid) {
     97   uint64_t start = NanoTime();
     98 
     99   siginfo_t si;
    100   while (ptrace(PTRACE_GETSIGINFO, pid, 0, &si) < 0 && (errno == EINTR || errno == ESRCH)) {
    101     if ((NanoTime() - start) > NS_PER_SEC) {
    102       printf("The process did not get to a stopping point in 1 second.\n");
    103       break;
    104     }
    105     usleep(US_PER_MSEC);
    106   }
    107 }
    108 
    109 bool ReadyLevelBacktrace(Backtrace* backtrace) {
    110   // See if test_level_four is in the backtrace.
    111   bool found = false;
    112   for (Backtrace::const_iterator it = backtrace->begin(); it != backtrace->end(); ++it) {
    113     if (it->func_name == "test_level_four") {
    114       found = true;
    115       break;
    116     }
    117   }
    118 
    119   return found;
    120 }
    121 
    122 void VerifyLevelDump(Backtrace* backtrace) {
    123   ASSERT_GT(backtrace->NumFrames(), static_cast<size_t>(0));
    124   ASSERT_LT(backtrace->NumFrames(), static_cast<size_t>(MAX_BACKTRACE_FRAMES));
    125 
    126   // Look through the frames starting at the highest to find the
    127   // frame we want.
    128   size_t frame_num = 0;
    129   for (size_t i = backtrace->NumFrames()-1; i > 2; i--) {
    130     if (backtrace->GetFrame(i)->func_name == "test_level_one") {
    131       frame_num = i;
    132       break;
    133     }
    134   }
    135   ASSERT_LT(static_cast<size_t>(0), frame_num);
    136   ASSERT_LE(static_cast<size_t>(3), frame_num);
    137 
    138   ASSERT_EQ(backtrace->GetFrame(frame_num)->func_name, "test_level_one");
    139   ASSERT_EQ(backtrace->GetFrame(frame_num-1)->func_name, "test_level_two");
    140   ASSERT_EQ(backtrace->GetFrame(frame_num-2)->func_name, "test_level_three");
    141   ASSERT_EQ(backtrace->GetFrame(frame_num-3)->func_name, "test_level_four");
    142 }
    143 
    144 void VerifyLevelBacktrace(void*) {
    145   UniquePtr<Backtrace> backtrace(
    146       Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD));
    147   ASSERT_TRUE(backtrace.get() != NULL);
    148   ASSERT_TRUE(backtrace->Unwind(0));
    149 
    150   VerifyLevelDump(backtrace.get());
    151 }
    152 
    153 bool ReadyMaxBacktrace(Backtrace* backtrace) {
    154   return (backtrace->NumFrames() == MAX_BACKTRACE_FRAMES);
    155 }
    156 
    157 void VerifyMaxDump(Backtrace* backtrace) {
    158   ASSERT_EQ(backtrace->NumFrames(), static_cast<size_t>(MAX_BACKTRACE_FRAMES));
    159   // Verify that the last frame is our recursive call.
    160   ASSERT_EQ(backtrace->GetFrame(MAX_BACKTRACE_FRAMES-1)->func_name,
    161             "test_recursive_call");
    162 }
    163 
    164 void VerifyMaxBacktrace(void*) {
    165   UniquePtr<Backtrace> backtrace(
    166       Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD));
    167   ASSERT_TRUE(backtrace.get() != NULL);
    168   ASSERT_TRUE(backtrace->Unwind(0));
    169 
    170   VerifyMaxDump(backtrace.get());
    171 }
    172 
    173 void ThreadSetState(void* data) {
    174   thread_t* thread = reinterpret_cast<thread_t*>(data);
    175   android_atomic_acquire_store(1, &thread->state);
    176   volatile int i = 0;
    177   while (thread->state) {
    178     i++;
    179   }
    180 }
    181 
    182 void VerifyThreadTest(pid_t tid, void (*VerifyFunc)(Backtrace*)) {
    183   UniquePtr<Backtrace> backtrace(Backtrace::Create(getpid(), tid));
    184   ASSERT_TRUE(backtrace.get() != NULL);
    185   ASSERT_TRUE(backtrace->Unwind(0));
    186 
    187   VerifyFunc(backtrace.get());
    188 }
    189 
    190 bool WaitForNonZero(int32_t* value, uint64_t seconds) {
    191   uint64_t start = NanoTime();
    192   do {
    193     if (android_atomic_acquire_load(value)) {
    194       return true;
    195     }
    196   } while ((NanoTime() - start) < seconds * NS_PER_SEC);
    197   return false;
    198 }
    199 
    200 TEST(libbacktrace, local_trace) {
    201   ASSERT_NE(test_level_one(1, 2, 3, 4, VerifyLevelBacktrace, NULL), 0);
    202 }
    203 
    204 void VerifyIgnoreFrames(
    205     Backtrace* bt_all, Backtrace* bt_ign1,
    206     Backtrace* bt_ign2, const char* cur_proc) {
    207   EXPECT_EQ(bt_all->NumFrames(), bt_ign1->NumFrames() + 1);
    208   EXPECT_EQ(bt_all->NumFrames(), bt_ign2->NumFrames() + 2);
    209 
    210   // Check all of the frames are the same > the current frame.
    211   bool check = (cur_proc == NULL);
    212   for (size_t i = 0; i < bt_ign2->NumFrames(); i++) {
    213     if (check) {
    214       EXPECT_EQ(bt_ign2->GetFrame(i)->pc, bt_ign1->GetFrame(i+1)->pc);
    215       EXPECT_EQ(bt_ign2->GetFrame(i)->sp, bt_ign1->GetFrame(i+1)->sp);
    216       EXPECT_EQ(bt_ign2->GetFrame(i)->stack_size, bt_ign1->GetFrame(i+1)->stack_size);
    217 
    218       EXPECT_EQ(bt_ign2->GetFrame(i)->pc, bt_all->GetFrame(i+2)->pc);
    219       EXPECT_EQ(bt_ign2->GetFrame(i)->sp, bt_all->GetFrame(i+2)->sp);
    220       EXPECT_EQ(bt_ign2->GetFrame(i)->stack_size, bt_all->GetFrame(i+2)->stack_size);
    221     }
    222     if (!check && bt_ign2->GetFrame(i)->func_name == cur_proc) {
    223       check = true;
    224     }
    225   }
    226 }
    227 
    228 void VerifyLevelIgnoreFrames(void*) {
    229   UniquePtr<Backtrace> all(
    230       Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD));
    231   ASSERT_TRUE(all.get() != NULL);
    232   ASSERT_TRUE(all->Unwind(0));
    233 
    234   UniquePtr<Backtrace> ign1(
    235       Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD));
    236   ASSERT_TRUE(ign1.get() != NULL);
    237   ASSERT_TRUE(ign1->Unwind(1));
    238 
    239   UniquePtr<Backtrace> ign2(
    240       Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD));
    241   ASSERT_TRUE(ign2.get() != NULL);
    242   ASSERT_TRUE(ign2->Unwind(2));
    243 
    244   VerifyIgnoreFrames(all.get(), ign1.get(), ign2.get(), "VerifyLevelIgnoreFrames");
    245 }
    246 
    247 TEST(libbacktrace, local_trace_ignore_frames) {
    248   ASSERT_NE(test_level_one(1, 2, 3, 4, VerifyLevelIgnoreFrames, NULL), 0);
    249 }
    250 
    251 TEST(libbacktrace, local_max_trace) {
    252   ASSERT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, VerifyMaxBacktrace, NULL), 0);
    253 }
    254 
    255 void VerifyProcTest(pid_t pid, pid_t tid, bool share_map,
    256                     bool (*ReadyFunc)(Backtrace*),
    257                     void (*VerifyFunc)(Backtrace*)) {
    258   pid_t ptrace_tid;
    259   if (tid < 0) {
    260     ptrace_tid = pid;
    261   } else {
    262     ptrace_tid = tid;
    263   }
    264   uint64_t start = NanoTime();
    265   bool verified = false;
    266   do {
    267     usleep(US_PER_MSEC);
    268     if (ptrace(PTRACE_ATTACH, ptrace_tid, 0, 0) == 0) {
    269       // Wait for the process to get to a stopping point.
    270       WaitForStop(ptrace_tid);
    271 
    272       UniquePtr<BacktraceMap> map;
    273       if (share_map) {
    274         map.reset(BacktraceMap::Create(pid));
    275       }
    276       UniquePtr<Backtrace> backtrace(Backtrace::Create(pid, tid, map.get()));
    277       ASSERT_TRUE(backtrace->Unwind(0));
    278       ASSERT_TRUE(backtrace.get() != NULL);
    279       if (ReadyFunc(backtrace.get())) {
    280         VerifyFunc(backtrace.get());
    281         verified = true;
    282       }
    283 
    284       ASSERT_TRUE(ptrace(PTRACE_DETACH, ptrace_tid, 0, 0) == 0);
    285     }
    286     // If 5 seconds have passed, then we are done.
    287   } while (!verified && (NanoTime() - start) <= 5 * NS_PER_SEC);
    288   ASSERT_TRUE(verified);
    289 }
    290 
    291 TEST(libbacktrace, ptrace_trace) {
    292   pid_t pid;
    293   if ((pid = fork()) == 0) {
    294     ASSERT_NE(test_level_one(1, 2, 3, 4, NULL, NULL), 0);
    295     _exit(1);
    296   }
    297   VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, false, ReadyLevelBacktrace, VerifyLevelDump);
    298 
    299   kill(pid, SIGKILL);
    300   int status;
    301   ASSERT_EQ(waitpid(pid, &status, 0), pid);
    302 }
    303 
    304 TEST(libbacktrace, ptrace_trace_shared_map) {
    305   pid_t pid;
    306   if ((pid = fork()) == 0) {
    307     ASSERT_NE(test_level_one(1, 2, 3, 4, NULL, NULL), 0);
    308     _exit(1);
    309   }
    310 
    311   VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, true, ReadyLevelBacktrace, VerifyLevelDump);
    312 
    313   kill(pid, SIGKILL);
    314   int status;
    315   ASSERT_EQ(waitpid(pid, &status, 0), pid);
    316 }
    317 
    318 TEST(libbacktrace, ptrace_max_trace) {
    319   pid_t pid;
    320   if ((pid = fork()) == 0) {
    321     ASSERT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, NULL, NULL), 0);
    322     _exit(1);
    323   }
    324   VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, false, ReadyMaxBacktrace, VerifyMaxDump);
    325 
    326   kill(pid, SIGKILL);
    327   int status;
    328   ASSERT_EQ(waitpid(pid, &status, 0), pid);
    329 }
    330 
    331 void VerifyProcessIgnoreFrames(Backtrace* bt_all) {
    332   UniquePtr<Backtrace> ign1(Backtrace::Create(bt_all->Pid(), BACKTRACE_CURRENT_THREAD));
    333   ASSERT_TRUE(ign1.get() != NULL);
    334   ASSERT_TRUE(ign1->Unwind(1));
    335 
    336   UniquePtr<Backtrace> ign2(Backtrace::Create(bt_all->Pid(), BACKTRACE_CURRENT_THREAD));
    337   ASSERT_TRUE(ign2.get() != NULL);
    338   ASSERT_TRUE(ign2->Unwind(2));
    339 
    340   VerifyIgnoreFrames(bt_all, ign1.get(), ign2.get(), NULL);
    341 }
    342 
    343 TEST(libbacktrace, ptrace_ignore_frames) {
    344   pid_t pid;
    345   if ((pid = fork()) == 0) {
    346     ASSERT_NE(test_level_one(1, 2, 3, 4, NULL, NULL), 0);
    347     _exit(1);
    348   }
    349   VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, false, ReadyLevelBacktrace, VerifyProcessIgnoreFrames);
    350 
    351   kill(pid, SIGKILL);
    352   int status;
    353   ASSERT_EQ(waitpid(pid, &status, 0), pid);
    354 }
    355 
    356 // Create a process with multiple threads and dump all of the threads.
    357 void* PtraceThreadLevelRun(void*) {
    358   EXPECT_NE(test_level_one(1, 2, 3, 4, NULL, NULL), 0);
    359   return NULL;
    360 }
    361 
    362 void GetThreads(pid_t pid, std::vector<pid_t>* threads) {
    363   // Get the list of tasks.
    364   char task_path[128];
    365   snprintf(task_path, sizeof(task_path), "/proc/%d/task", pid);
    366 
    367   DIR* tasks_dir = opendir(task_path);
    368   ASSERT_TRUE(tasks_dir != NULL);
    369   struct dirent* entry;
    370   while ((entry = readdir(tasks_dir)) != NULL) {
    371     char* end;
    372     pid_t tid = strtoul(entry->d_name, &end, 10);
    373     if (*end == '\0') {
    374       threads->push_back(tid);
    375     }
    376   }
    377   closedir(tasks_dir);
    378 }
    379 
    380 TEST(libbacktrace, ptrace_threads) {
    381   pid_t pid;
    382   if ((pid = fork()) == 0) {
    383     for (size_t i = 0; i < NUM_PTRACE_THREADS; i++) {
    384       pthread_attr_t attr;
    385       pthread_attr_init(&attr);
    386       pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
    387 
    388       pthread_t thread;
    389       ASSERT_TRUE(pthread_create(&thread, &attr, PtraceThreadLevelRun, NULL) == 0);
    390     }
    391     ASSERT_NE(test_level_one(1, 2, 3, 4, NULL, NULL), 0);
    392     _exit(1);
    393   }
    394 
    395   // Check to see that all of the threads are running before unwinding.
    396   std::vector<pid_t> threads;
    397   uint64_t start = NanoTime();
    398   do {
    399     usleep(US_PER_MSEC);
    400     threads.clear();
    401     GetThreads(pid, &threads);
    402   } while ((threads.size() != NUM_PTRACE_THREADS + 1) &&
    403       ((NanoTime() - start) <= 5 * NS_PER_SEC));
    404   ASSERT_EQ(threads.size(), static_cast<size_t>(NUM_PTRACE_THREADS + 1));
    405 
    406   ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0);
    407   WaitForStop(pid);
    408   for (std::vector<int>::const_iterator it = threads.begin(); it != threads.end(); ++it) {
    409     // Skip the current forked process, we only care about the threads.
    410     if (pid == *it) {
    411       continue;
    412     }
    413     VerifyProcTest(pid, *it, false, ReadyLevelBacktrace, VerifyLevelDump);
    414   }
    415   ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0);
    416 
    417   kill(pid, SIGKILL);
    418   int status;
    419   ASSERT_EQ(waitpid(pid, &status, 0), pid);
    420 }
    421 
    422 void VerifyLevelThread(void*) {
    423   UniquePtr<Backtrace> backtrace(Backtrace::Create(getpid(), gettid()));
    424   ASSERT_TRUE(backtrace.get() != NULL);
    425   ASSERT_TRUE(backtrace->Unwind(0));
    426 
    427   VerifyLevelDump(backtrace.get());
    428 }
    429 
    430 TEST(libbacktrace, thread_current_level) {
    431   ASSERT_NE(test_level_one(1, 2, 3, 4, VerifyLevelThread, NULL), 0);
    432 }
    433 
    434 void VerifyMaxThread(void*) {
    435   UniquePtr<Backtrace> backtrace(Backtrace::Create(getpid(), gettid()));
    436   ASSERT_TRUE(backtrace.get() != NULL);
    437   ASSERT_TRUE(backtrace->Unwind(0));
    438 
    439   VerifyMaxDump(backtrace.get());
    440 }
    441 
    442 TEST(libbacktrace, thread_current_max) {
    443   ASSERT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, VerifyMaxThread, NULL), 0);
    444 }
    445 
    446 void* ThreadLevelRun(void* data) {
    447   thread_t* thread = reinterpret_cast<thread_t*>(data);
    448 
    449   thread->tid = gettid();
    450   EXPECT_NE(test_level_one(1, 2, 3, 4, ThreadSetState, data), 0);
    451   return NULL;
    452 }
    453 
    454 TEST(libbacktrace, thread_level_trace) {
    455   pthread_attr_t attr;
    456   pthread_attr_init(&attr);
    457   pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
    458 
    459   thread_t thread_data = { 0, 0, 0 };
    460   pthread_t thread;
    461   ASSERT_TRUE(pthread_create(&thread, &attr, ThreadLevelRun, &thread_data) == 0);
    462 
    463   // Wait up to 2 seconds for the tid to be set.
    464   ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2));
    465 
    466   // Make sure that the thread signal used is not visible when compiled for
    467   // the target.
    468 #if !defined(__GLIBC__)
    469   ASSERT_LT(THREAD_SIGNAL, SIGRTMIN);
    470 #endif
    471 
    472   // Save the current signal action and make sure it is restored afterwards.
    473   struct sigaction cur_action;
    474   ASSERT_TRUE(sigaction(THREAD_SIGNAL, NULL, &cur_action) == 0);
    475 
    476   UniquePtr<Backtrace> backtrace(Backtrace::Create(getpid(), thread_data.tid));
    477   ASSERT_TRUE(backtrace.get() != NULL);
    478   ASSERT_TRUE(backtrace->Unwind(0));
    479 
    480   VerifyLevelDump(backtrace.get());
    481 
    482   // Tell the thread to exit its infinite loop.
    483   android_atomic_acquire_store(0, &thread_data.state);
    484 
    485   // Verify that the old action was restored.
    486   struct sigaction new_action;
    487   ASSERT_TRUE(sigaction(THREAD_SIGNAL, NULL, &new_action) == 0);
    488   EXPECT_EQ(cur_action.sa_sigaction, new_action.sa_sigaction);
    489   // The SA_RESTORER flag gets set behind our back, so a direct comparison
    490   // doesn't work unless we mask the value off. Mips doesn't have this
    491   // flag, so skip this on that platform.
    492 #ifdef SA_RESTORER
    493   cur_action.sa_flags &= ~SA_RESTORER;
    494   new_action.sa_flags &= ~SA_RESTORER;
    495 #endif
    496   EXPECT_EQ(cur_action.sa_flags, new_action.sa_flags);
    497 }
    498 
    499 TEST(libbacktrace, thread_ignore_frames) {
    500   pthread_attr_t attr;
    501   pthread_attr_init(&attr);
    502   pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
    503 
    504   thread_t thread_data = { 0, 0, 0 };
    505   pthread_t thread;
    506   ASSERT_TRUE(pthread_create(&thread, &attr, ThreadLevelRun, &thread_data) == 0);
    507 
    508   // Wait up to 2 seconds for the tid to be set.
    509   ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2));
    510 
    511   UniquePtr<Backtrace> all(Backtrace::Create(getpid(), thread_data.tid));
    512   ASSERT_TRUE(all.get() != NULL);
    513   ASSERT_TRUE(all->Unwind(0));
    514 
    515   UniquePtr<Backtrace> ign1(Backtrace::Create(getpid(), thread_data.tid));
    516   ASSERT_TRUE(ign1.get() != NULL);
    517   ASSERT_TRUE(ign1->Unwind(1));
    518 
    519   UniquePtr<Backtrace> ign2(Backtrace::Create(getpid(), thread_data.tid));
    520   ASSERT_TRUE(ign2.get() != NULL);
    521   ASSERT_TRUE(ign2->Unwind(2));
    522 
    523   VerifyIgnoreFrames(all.get(), ign1.get(), ign2.get(), NULL);
    524 
    525   // Tell the thread to exit its infinite loop.
    526   android_atomic_acquire_store(0, &thread_data.state);
    527 }
    528 
    529 void* ThreadMaxRun(void* data) {
    530   thread_t* thread = reinterpret_cast<thread_t*>(data);
    531 
    532   thread->tid = gettid();
    533   EXPECT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, ThreadSetState, data), 0);
    534   return NULL;
    535 }
    536 
    537 TEST(libbacktrace, thread_max_trace) {
    538   pthread_attr_t attr;
    539   pthread_attr_init(&attr);
    540   pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
    541 
    542   thread_t thread_data = { 0, 0, 0 };
    543   pthread_t thread;
    544   ASSERT_TRUE(pthread_create(&thread, &attr, ThreadMaxRun, &thread_data) == 0);
    545 
    546   // Wait for the tid to be set.
    547   ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2));
    548 
    549   UniquePtr<Backtrace> backtrace(Backtrace::Create(getpid(), thread_data.tid));
    550   ASSERT_TRUE(backtrace.get() != NULL);
    551   ASSERT_TRUE(backtrace->Unwind(0));
    552 
    553   VerifyMaxDump(backtrace.get());
    554 
    555   // Tell the thread to exit its infinite loop.
    556   android_atomic_acquire_store(0, &thread_data.state);
    557 }
    558 
    559 void* ThreadDump(void* data) {
    560   dump_thread_t* dump = reinterpret_cast<dump_thread_t*>(data);
    561   while (true) {
    562     if (android_atomic_acquire_load(dump->now)) {
    563       break;
    564     }
    565   }
    566 
    567   // The status of the actual unwind will be checked elsewhere.
    568   dump->backtrace = Backtrace::Create(getpid(), dump->thread.tid);
    569   dump->backtrace->Unwind(0);
    570 
    571   android_atomic_acquire_store(1, &dump->done);
    572 
    573   return NULL;
    574 }
    575 
    576 TEST(libbacktrace, thread_multiple_dump) {
    577   // Dump NUM_THREADS simultaneously.
    578   std::vector<thread_t> runners(NUM_THREADS);
    579   std::vector<dump_thread_t> dumpers(NUM_THREADS);
    580 
    581   pthread_attr_t attr;
    582   pthread_attr_init(&attr);
    583   pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
    584   for (size_t i = 0; i < NUM_THREADS; i++) {
    585     // Launch the runners, they will spin in hard loops doing nothing.
    586     runners[i].tid = 0;
    587     runners[i].state = 0;
    588     ASSERT_TRUE(pthread_create(&runners[i].threadId, &attr, ThreadMaxRun, &runners[i]) == 0);
    589   }
    590 
    591   // Wait for tids to be set.
    592   for (std::vector<thread_t>::iterator it = runners.begin(); it != runners.end(); ++it) {
    593     ASSERT_TRUE(WaitForNonZero(&it->state, 30));
    594   }
    595 
    596   // Start all of the dumpers at once, they will spin until they are signalled
    597   // to begin their dump run.
    598   int32_t dump_now = 0;
    599   for (size_t i = 0; i < NUM_THREADS; i++) {
    600     dumpers[i].thread.tid = runners[i].tid;
    601     dumpers[i].thread.state = 0;
    602     dumpers[i].done = 0;
    603     dumpers[i].now = &dump_now;
    604 
    605     ASSERT_TRUE(pthread_create(&dumpers[i].thread.threadId, &attr, ThreadDump, &dumpers[i]) == 0);
    606   }
    607 
    608   // Start all of the dumpers going at once.
    609   android_atomic_acquire_store(1, &dump_now);
    610 
    611   for (size_t i = 0; i < NUM_THREADS; i++) {
    612     ASSERT_TRUE(WaitForNonZero(&dumpers[i].done, 30));
    613 
    614     // Tell the runner thread to exit its infinite loop.
    615     android_atomic_acquire_store(0, &runners[i].state);
    616 
    617     ASSERT_TRUE(dumpers[i].backtrace != NULL);
    618     VerifyMaxDump(dumpers[i].backtrace);
    619 
    620     delete dumpers[i].backtrace;
    621     dumpers[i].backtrace = NULL;
    622   }
    623 }
    624 
    625 TEST(libbacktrace, thread_multiple_dump_same_thread) {
    626   pthread_attr_t attr;
    627   pthread_attr_init(&attr);
    628   pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
    629   thread_t runner;
    630   runner.tid = 0;
    631   runner.state = 0;
    632   ASSERT_TRUE(pthread_create(&runner.threadId, &attr, ThreadMaxRun, &runner) == 0);
    633 
    634   // Wait for tids to be set.
    635   ASSERT_TRUE(WaitForNonZero(&runner.state, 30));
    636 
    637   // Start all of the dumpers at once, they will spin until they are signalled
    638   // to begin their dump run.
    639   int32_t dump_now = 0;
    640   // Dump the same thread NUM_THREADS simultaneously.
    641   std::vector<dump_thread_t> dumpers(NUM_THREADS);
    642   for (size_t i = 0; i < NUM_THREADS; i++) {
    643     dumpers[i].thread.tid = runner.tid;
    644     dumpers[i].thread.state = 0;
    645     dumpers[i].done = 0;
    646     dumpers[i].now = &dump_now;
    647 
    648     ASSERT_TRUE(pthread_create(&dumpers[i].thread.threadId, &attr, ThreadDump, &dumpers[i]) == 0);
    649   }
    650 
    651   // Start all of the dumpers going at once.
    652   android_atomic_acquire_store(1, &dump_now);
    653 
    654   for (size_t i = 0; i < NUM_THREADS; i++) {
    655     ASSERT_TRUE(WaitForNonZero(&dumpers[i].done, 30));
    656 
    657     ASSERT_TRUE(dumpers[i].backtrace != NULL);
    658     VerifyMaxDump(dumpers[i].backtrace);
    659 
    660     delete dumpers[i].backtrace;
    661     dumpers[i].backtrace = NULL;
    662   }
    663 
    664   // Tell the runner thread to exit its infinite loop.
    665   android_atomic_acquire_store(0, &runner.state);
    666 }
    667 
    668 // This test is for UnwindMaps that should share the same map cursor when
    669 // multiple maps are created for the current process at the same time.
    670 TEST(libbacktrace, simultaneous_maps) {
    671   BacktraceMap* map1 = BacktraceMap::Create(getpid());
    672   BacktraceMap* map2 = BacktraceMap::Create(getpid());
    673   BacktraceMap* map3 = BacktraceMap::Create(getpid());
    674 
    675   Backtrace* back1 = Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD, map1);
    676   EXPECT_TRUE(back1->Unwind(0));
    677   delete back1;
    678   delete map1;
    679 
    680   Backtrace* back2 = Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD, map2);
    681   EXPECT_TRUE(back2->Unwind(0));
    682   delete back2;
    683   delete map2;
    684 
    685   Backtrace* back3 = Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD, map3);
    686   EXPECT_TRUE(back3->Unwind(0));
    687   delete back3;
    688   delete map3;
    689 }
    690 
    691 TEST(libbacktrace, format_test) {
    692   UniquePtr<Backtrace> backtrace(Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD));
    693   ASSERT_TRUE(backtrace.get() != NULL);
    694 
    695   backtrace_frame_data_t frame;
    696   frame.num = 1;
    697   frame.pc = 2;
    698   frame.sp = 0;
    699   frame.stack_size = 0;
    700   frame.map = NULL;
    701   frame.func_offset = 0;
    702 
    703   backtrace_map_t map;
    704   map.start = 0;
    705   map.end = 0;
    706 
    707   // Check no map set.
    708   frame.num = 1;
    709 #if defined(__LP64__)
    710   EXPECT_EQ("#01 pc 0000000000000002  <unknown>",
    711 #else
    712   EXPECT_EQ("#01 pc 00000002  <unknown>",
    713 #endif
    714             backtrace->FormatFrameData(&frame));
    715 
    716   // Check map name empty, but exists.
    717   frame.map = &map;
    718   map.start = 1;
    719 #if defined(__LP64__)
    720   EXPECT_EQ("#01 pc 0000000000000001  <unknown>",
    721 #else
    722   EXPECT_EQ("#01 pc 00000001  <unknown>",
    723 #endif
    724             backtrace->FormatFrameData(&frame));
    725 
    726 
    727   // Check relative pc is set and map name is set.
    728   frame.pc = 0x12345679;
    729   frame.map = &map;
    730   map.name = "MapFake";
    731   map.start =  1;
    732 #if defined(__LP64__)
    733   EXPECT_EQ("#01 pc 0000000012345678  MapFake",
    734 #else
    735   EXPECT_EQ("#01 pc 12345678  MapFake",
    736 #endif
    737             backtrace->FormatFrameData(&frame));
    738 
    739   // Check func_name is set, but no func offset.
    740   frame.func_name = "ProcFake";
    741 #if defined(__LP64__)
    742   EXPECT_EQ("#01 pc 0000000012345678  MapFake (ProcFake)",
    743 #else
    744   EXPECT_EQ("#01 pc 12345678  MapFake (ProcFake)",
    745 #endif
    746             backtrace->FormatFrameData(&frame));
    747 
    748   // Check func_name is set, and func offset is non-zero.
    749   frame.func_offset = 645;
    750 #if defined(__LP64__)
    751   EXPECT_EQ("#01 pc 0000000012345678  MapFake (ProcFake+645)",
    752 #else
    753   EXPECT_EQ("#01 pc 12345678  MapFake (ProcFake+645)",
    754 #endif
    755             backtrace->FormatFrameData(&frame));
    756 }
    757 
    758 struct map_test_t {
    759   uintptr_t start;
    760   uintptr_t end;
    761 };
    762 
    763 bool map_sort(map_test_t i, map_test_t j) {
    764   return i.start < j.start;
    765 }
    766 
    767 static void VerifyMap(pid_t pid) {
    768   char buffer[4096];
    769   snprintf(buffer, sizeof(buffer), "/proc/%d/maps", pid);
    770 
    771   FILE* map_file = fopen(buffer, "r");
    772   ASSERT_TRUE(map_file != NULL);
    773   std::vector<map_test_t> test_maps;
    774   while (fgets(buffer, sizeof(buffer), map_file)) {
    775     map_test_t map;
    776     ASSERT_EQ(2, sscanf(buffer, "%" SCNxPTR "-%" SCNxPTR " ", &map.start, &map.end));
    777     test_maps.push_back(map);
    778   }
    779   fclose(map_file);
    780   std::sort(test_maps.begin(), test_maps.end(), map_sort);
    781 
    782   UniquePtr<BacktraceMap> map(BacktraceMap::Create(pid));
    783 
    784   // Basic test that verifies that the map is in the expected order.
    785   std::vector<map_test_t>::const_iterator test_it = test_maps.begin();
    786   for (BacktraceMap::const_iterator it = map->begin(); it != map->end(); ++it) {
    787     ASSERT_TRUE(test_it != test_maps.end());
    788     ASSERT_EQ(test_it->start, it->start);
    789     ASSERT_EQ(test_it->end, it->end);
    790     ++test_it;
    791   }
    792   ASSERT_TRUE(test_it == test_maps.end());
    793 }
    794 
    795 TEST(libbacktrace, verify_map_remote) {
    796   pid_t pid;
    797 
    798   if ((pid = fork()) == 0) {
    799     while (true) {
    800     }
    801     _exit(0);
    802   }
    803   ASSERT_LT(0, pid);
    804 
    805   ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0);
    806 
    807   // Wait for the process to get to a stopping point.
    808   WaitForStop(pid);
    809 
    810   // The maps should match exactly since the forked process has been paused.
    811   VerifyMap(pid);
    812 
    813   ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0);
    814 
    815   kill(pid, SIGKILL);
    816   ASSERT_EQ(waitpid(pid, NULL, 0), pid);
    817 }
    818 
    819 #if defined(ENABLE_PSS_TESTS)
    820 #include "GetPss.h"
    821 
    822 #define MAX_LEAK_BYTES 32*1024UL
    823 
    824 static void CheckForLeak(pid_t pid, pid_t tid) {
    825   // Do a few runs to get the PSS stable.
    826   for (size_t i = 0; i < 100; i++) {
    827     Backtrace* backtrace = Backtrace::Create(pid, tid);
    828     ASSERT_TRUE(backtrace != NULL);
    829     ASSERT_TRUE(backtrace->Unwind(0));
    830     delete backtrace;
    831   }
    832   size_t stable_pss = GetPssBytes();
    833 
    834   // Loop enough that even a small leak should be detectable.
    835   for (size_t i = 0; i < 4096; i++) {
    836     Backtrace* backtrace = Backtrace::Create(pid, tid);
    837     ASSERT_TRUE(backtrace != NULL);
    838     ASSERT_TRUE(backtrace->Unwind(0));
    839     delete backtrace;
    840   }
    841   size_t new_pss = GetPssBytes();
    842   size_t abs_diff = (new_pss > stable_pss) ? new_pss - stable_pss : stable_pss - new_pss;
    843   // As long as the new pss is within a certain amount, consider everything okay.
    844   ASSERT_LE(abs_diff, MAX_LEAK_BYTES);
    845 }
    846 
    847 TEST(libbacktrace, check_for_leak_local) {
    848   CheckForLeak(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD);
    849 }
    850 
    851 TEST(libbacktrace, check_for_leak_local_thread) {
    852   thread_t thread_data = { 0, 0, 0 };
    853   pthread_t thread;
    854   ASSERT_TRUE(pthread_create(&thread, NULL, ThreadLevelRun, &thread_data) == 0);
    855 
    856   // Wait up to 2 seconds for the tid to be set.
    857   ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2));
    858 
    859   CheckForLeak(BACKTRACE_CURRENT_PROCESS, thread_data.tid);
    860 
    861   // Tell the thread to exit its infinite loop.
    862   android_atomic_acquire_store(0, &thread_data.state);
    863 
    864   ASSERT_TRUE(pthread_join(thread, NULL) == 0);
    865 }
    866 
    867 TEST(libbacktrace, check_for_leak_remote) {
    868   pid_t pid;
    869 
    870   if ((pid = fork()) == 0) {
    871     while (true) {
    872     }
    873     _exit(0);
    874   }
    875   ASSERT_LT(0, pid);
    876 
    877   ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0);
    878 
    879   // Wait for the process to get to a stopping point.
    880   WaitForStop(pid);
    881 
    882   CheckForLeak(pid, BACKTRACE_CURRENT_THREAD);
    883 
    884   ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0);
    885 
    886   kill(pid, SIGKILL);
    887   ASSERT_EQ(waitpid(pid, NULL, 0), pid);
    888 }
    889 #endif
    890