Home | History | Annotate | Download | only in tests
      1 /*
      2  * Copyright (C) 2012 The Android Open Source Project
      3  *
      4  * Licensed under the Apache License, Version 2.0 (the "License");
      5  * you may not use this file except in compliance with the License.
      6  * You may obtain a copy of the License at
      7  *
      8  *      http://www.apache.org/licenses/LICENSE-2.0
      9  *
     10  * Unless required by applicable law or agreed to in writing, software
     11  * distributed under the License is distributed on an "AS IS" BASIS,
     12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     13  * See the License for the specific language governing permissions and
     14  * limitations under the License.
     15  */
     16 
     17 #include <gtest/gtest.h>
     18 
     19 #include <errno.h>
     20 #include <inttypes.h>
     21 #include <limits.h>
     22 #include <malloc.h>
     23 #include <pthread.h>
     24 #include <signal.h>
     25 #include <sys/mman.h>
     26 #include <sys/syscall.h>
     27 #include <time.h>
     28 #include <unistd.h>
     29 
     30 #include "private/ScopeGuard.h"
     31 #include "ScopedSignalHandler.h"
     32 
     33 TEST(pthread, pthread_key_create) {
     34   pthread_key_t key;
     35   ASSERT_EQ(0, pthread_key_create(&key, NULL));
     36   ASSERT_EQ(0, pthread_key_delete(key));
     37   // Can't delete a key that's already been deleted.
     38   ASSERT_EQ(EINVAL, pthread_key_delete(key));
     39 }
     40 
     41 TEST(pthread, pthread_key_create_lots) {
     42 #if defined(__BIONIC__) // glibc uses keys internally that its sysconf value doesn't account for.
     43   // POSIX says PTHREAD_KEYS_MAX should be at least 128.
     44   ASSERT_GE(PTHREAD_KEYS_MAX, 128);
     45 
     46   int sysconf_max = sysconf(_SC_THREAD_KEYS_MAX);
     47 
     48   // sysconf shouldn't return a smaller value.
     49   ASSERT_GE(sysconf_max, PTHREAD_KEYS_MAX);
     50 
     51   // We can allocate _SC_THREAD_KEYS_MAX keys.
     52   sysconf_max -= 2; // (Except that gtest takes two for itself.)
     53   std::vector<pthread_key_t> keys;
     54   for (int i = 0; i < sysconf_max; ++i) {
     55     pthread_key_t key;
     56     // If this fails, it's likely that GLOBAL_INIT_THREAD_LOCAL_BUFFER_COUNT is wrong.
     57     ASSERT_EQ(0, pthread_key_create(&key, NULL)) << i << " of " << sysconf_max;
     58     keys.push_back(key);
     59   }
     60 
     61   // ...and that really is the maximum.
     62   pthread_key_t key;
     63   ASSERT_EQ(EAGAIN, pthread_key_create(&key, NULL));
     64 
     65   // (Don't leak all those keys!)
     66   for (size_t i = 0; i < keys.size(); ++i) {
     67     ASSERT_EQ(0, pthread_key_delete(keys[i]));
     68   }
     69 #else // __BIONIC__
     70   GTEST_LOG_(INFO) << "This test does nothing.\n";
     71 #endif // __BIONIC__
     72 }
     73 
     74 TEST(pthread, pthread_key_delete) {
     75   void* expected = reinterpret_cast<void*>(1234);
     76   pthread_key_t key;
     77   ASSERT_EQ(0, pthread_key_create(&key, NULL));
     78   ASSERT_EQ(0, pthread_setspecific(key, expected));
     79   ASSERT_EQ(expected, pthread_getspecific(key));
     80   ASSERT_EQ(0, pthread_key_delete(key));
     81   // After deletion, pthread_getspecific returns NULL.
     82   ASSERT_EQ(NULL, pthread_getspecific(key));
     83   // And you can't use pthread_setspecific with the deleted key.
     84   ASSERT_EQ(EINVAL, pthread_setspecific(key, expected));
     85 }
     86 
     87 TEST(pthread, pthread_key_fork) {
     88   void* expected = reinterpret_cast<void*>(1234);
     89   pthread_key_t key;
     90   ASSERT_EQ(0, pthread_key_create(&key, NULL));
     91   ASSERT_EQ(0, pthread_setspecific(key, expected));
     92   ASSERT_EQ(expected, pthread_getspecific(key));
     93 
     94   pid_t pid = fork();
     95   ASSERT_NE(-1, pid) << strerror(errno);
     96 
     97   if (pid == 0) {
     98     // The surviving thread inherits all the forking thread's TLS values...
     99     ASSERT_EQ(expected, pthread_getspecific(key));
    100     _exit(99);
    101   }
    102 
    103   int status;
    104   ASSERT_EQ(pid, waitpid(pid, &status, 0));
    105   ASSERT_TRUE(WIFEXITED(status));
    106   ASSERT_EQ(99, WEXITSTATUS(status));
    107 
    108   ASSERT_EQ(expected, pthread_getspecific(key));
    109 }
    110 
    111 static void* DirtyKeyFn(void* key) {
    112   return pthread_getspecific(*reinterpret_cast<pthread_key_t*>(key));
    113 }
    114 
    115 TEST(pthread, pthread_key_dirty) {
    116   pthread_key_t key;
    117   ASSERT_EQ(0, pthread_key_create(&key, NULL));
    118 
    119   size_t stack_size = 128 * 1024;
    120   void* stack = mmap(NULL, stack_size, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0);
    121   ASSERT_NE(MAP_FAILED, stack);
    122   memset(stack, 0xff, stack_size);
    123 
    124   pthread_attr_t attr;
    125   ASSERT_EQ(0, pthread_attr_init(&attr));
    126   ASSERT_EQ(0, pthread_attr_setstack(&attr, stack, stack_size));
    127 
    128   pthread_t t;
    129   ASSERT_EQ(0, pthread_create(&t, &attr, DirtyKeyFn, &key));
    130 
    131   void* result;
    132   ASSERT_EQ(0, pthread_join(t, &result));
    133   ASSERT_EQ(nullptr, result); // Not ~0!
    134 
    135   ASSERT_EQ(0, munmap(stack, stack_size));
    136 }
    137 
    138 static void* IdFn(void* arg) {
    139   return arg;
    140 }
    141 
    142 static void* SleepFn(void* arg) {
    143   sleep(reinterpret_cast<uintptr_t>(arg));
    144   return NULL;
    145 }
    146 
    147 static void* SpinFn(void* arg) {
    148   volatile bool* b = reinterpret_cast<volatile bool*>(arg);
    149   while (!*b) {
    150   }
    151   return NULL;
    152 }
    153 
    154 static void* JoinFn(void* arg) {
    155   return reinterpret_cast<void*>(pthread_join(reinterpret_cast<pthread_t>(arg), NULL));
    156 }
    157 
    158 static void AssertDetached(pthread_t t, bool is_detached) {
    159   pthread_attr_t attr;
    160   ASSERT_EQ(0, pthread_getattr_np(t, &attr));
    161   int detach_state;
    162   ASSERT_EQ(0, pthread_attr_getdetachstate(&attr, &detach_state));
    163   pthread_attr_destroy(&attr);
    164   ASSERT_EQ(is_detached, (detach_state == PTHREAD_CREATE_DETACHED));
    165 }
    166 
    167 static void MakeDeadThread(pthread_t& t) {
    168   ASSERT_EQ(0, pthread_create(&t, NULL, IdFn, NULL));
    169   void* result;
    170   ASSERT_EQ(0, pthread_join(t, &result));
    171 }
    172 
    173 TEST(pthread, pthread_create) {
    174   void* expected_result = reinterpret_cast<void*>(123);
    175   // Can we create a thread?
    176   pthread_t t;
    177   ASSERT_EQ(0, pthread_create(&t, NULL, IdFn, expected_result));
    178   // If we join, do we get the expected value back?
    179   void* result;
    180   ASSERT_EQ(0, pthread_join(t, &result));
    181   ASSERT_EQ(expected_result, result);
    182 }
    183 
    184 TEST(pthread, pthread_create_EAGAIN) {
    185   pthread_attr_t attributes;
    186   ASSERT_EQ(0, pthread_attr_init(&attributes));
    187   ASSERT_EQ(0, pthread_attr_setstacksize(&attributes, static_cast<size_t>(-1) & ~(getpagesize() - 1)));
    188 
    189   pthread_t t;
    190   ASSERT_EQ(EAGAIN, pthread_create(&t, &attributes, IdFn, NULL));
    191 }
    192 
    193 TEST(pthread, pthread_no_join_after_detach) {
    194   pthread_t t1;
    195   ASSERT_EQ(0, pthread_create(&t1, NULL, SleepFn, reinterpret_cast<void*>(5)));
    196 
    197   // After a pthread_detach...
    198   ASSERT_EQ(0, pthread_detach(t1));
    199   AssertDetached(t1, true);
    200 
    201   // ...pthread_join should fail.
    202   void* result;
    203   ASSERT_EQ(EINVAL, pthread_join(t1, &result));
    204 }
    205 
    206 TEST(pthread, pthread_no_op_detach_after_join) {
    207   bool done = false;
    208 
    209   pthread_t t1;
    210   ASSERT_EQ(0, pthread_create(&t1, NULL, SpinFn, &done));
    211 
    212   // If thread 2 is already waiting to join thread 1...
    213   pthread_t t2;
    214   ASSERT_EQ(0, pthread_create(&t2, NULL, JoinFn, reinterpret_cast<void*>(t1)));
    215 
    216   sleep(1); // (Give t2 a chance to call pthread_join.)
    217 
    218   // ...a call to pthread_detach on thread 1 will "succeed" (silently fail)...
    219   ASSERT_EQ(0, pthread_detach(t1));
    220   AssertDetached(t1, false);
    221 
    222   done = true;
    223 
    224   // ...but t2's join on t1 still goes ahead (which we can tell because our join on t2 finishes).
    225   void* join_result;
    226   ASSERT_EQ(0, pthread_join(t2, &join_result));
    227   ASSERT_EQ(0U, reinterpret_cast<uintptr_t>(join_result));
    228 }
    229 
    230 TEST(pthread, pthread_join_self) {
    231   void* result;
    232   ASSERT_EQ(EDEADLK, pthread_join(pthread_self(), &result));
    233 }
    234 
    235 struct TestBug37410 {
    236   pthread_t main_thread;
    237   pthread_mutex_t mutex;
    238 
    239   static void main() {
    240     TestBug37410 data;
    241     data.main_thread = pthread_self();
    242     ASSERT_EQ(0, pthread_mutex_init(&data.mutex, NULL));
    243     ASSERT_EQ(0, pthread_mutex_lock(&data.mutex));
    244 
    245     pthread_t t;
    246     ASSERT_EQ(0, pthread_create(&t, NULL, TestBug37410::thread_fn, reinterpret_cast<void*>(&data)));
    247 
    248     // Wait for the thread to be running...
    249     ASSERT_EQ(0, pthread_mutex_lock(&data.mutex));
    250     ASSERT_EQ(0, pthread_mutex_unlock(&data.mutex));
    251 
    252     // ...and exit.
    253     pthread_exit(NULL);
    254   }
    255 
    256  private:
    257   static void* thread_fn(void* arg) {
    258     TestBug37410* data = reinterpret_cast<TestBug37410*>(arg);
    259 
    260     // Let the main thread know we're running.
    261     pthread_mutex_unlock(&data->mutex);
    262 
    263     // And wait for the main thread to exit.
    264     pthread_join(data->main_thread, NULL);
    265 
    266     return NULL;
    267   }
    268 };
    269 
    270 // Even though this isn't really a death test, we have to say "DeathTest" here so gtest knows to
    271 // run this test (which exits normally) in its own process.
    272 TEST(pthread_DeathTest, pthread_bug_37410) {
    273   // http://code.google.com/p/android/issues/detail?id=37410
    274   ::testing::FLAGS_gtest_death_test_style = "threadsafe";
    275   ASSERT_EXIT(TestBug37410::main(), ::testing::ExitedWithCode(0), "");
    276 }
    277 
    278 static void* SignalHandlerFn(void* arg) {
    279   sigset_t wait_set;
    280   sigfillset(&wait_set);
    281   return reinterpret_cast<void*>(sigwait(&wait_set, reinterpret_cast<int*>(arg)));
    282 }
    283 
    284 TEST(pthread, pthread_sigmask) {
    285   // Check that SIGUSR1 isn't blocked.
    286   sigset_t original_set;
    287   sigemptyset(&original_set);
    288   ASSERT_EQ(0, pthread_sigmask(SIG_BLOCK, NULL, &original_set));
    289   ASSERT_FALSE(sigismember(&original_set, SIGUSR1));
    290 
    291   // Block SIGUSR1.
    292   sigset_t set;
    293   sigemptyset(&set);
    294   sigaddset(&set, SIGUSR1);
    295   ASSERT_EQ(0, pthread_sigmask(SIG_BLOCK, &set, NULL));
    296 
    297   // Check that SIGUSR1 is blocked.
    298   sigset_t final_set;
    299   sigemptyset(&final_set);
    300   ASSERT_EQ(0, pthread_sigmask(SIG_BLOCK, NULL, &final_set));
    301   ASSERT_TRUE(sigismember(&final_set, SIGUSR1));
    302   // ...and that sigprocmask agrees with pthread_sigmask.
    303   sigemptyset(&final_set);
    304   ASSERT_EQ(0, sigprocmask(SIG_BLOCK, NULL, &final_set));
    305   ASSERT_TRUE(sigismember(&final_set, SIGUSR1));
    306 
    307   // Spawn a thread that calls sigwait and tells us what it received.
    308   pthread_t signal_thread;
    309   int received_signal = -1;
    310   ASSERT_EQ(0, pthread_create(&signal_thread, NULL, SignalHandlerFn, &received_signal));
    311 
    312   // Send that thread SIGUSR1.
    313   pthread_kill(signal_thread, SIGUSR1);
    314 
    315   // See what it got.
    316   void* join_result;
    317   ASSERT_EQ(0, pthread_join(signal_thread, &join_result));
    318   ASSERT_EQ(SIGUSR1, received_signal);
    319   ASSERT_EQ(0U, reinterpret_cast<uintptr_t>(join_result));
    320 
    321   // Restore the original signal mask.
    322   ASSERT_EQ(0, pthread_sigmask(SIG_SETMASK, &original_set, NULL));
    323 }
    324 
    325 TEST(pthread, pthread_setname_np__too_long) {
    326 #if defined(__BIONIC__) // Not all build servers have a new enough glibc? TODO: remove when they're on gprecise.
    327   ASSERT_EQ(ERANGE, pthread_setname_np(pthread_self(), "this name is far too long for linux"));
    328 #else // __BIONIC__
    329   GTEST_LOG_(INFO) << "This test does nothing.\n";
    330 #endif // __BIONIC__
    331 }
    332 
    333 TEST(pthread, pthread_setname_np__self) {
    334 #if defined(__BIONIC__) // Not all build servers have a new enough glibc? TODO: remove when they're on gprecise.
    335   ASSERT_EQ(0, pthread_setname_np(pthread_self(), "short 1"));
    336 #else // __BIONIC__
    337   GTEST_LOG_(INFO) << "This test does nothing.\n";
    338 #endif // __BIONIC__
    339 }
    340 
    341 TEST(pthread, pthread_setname_np__other) {
    342 #if defined(__BIONIC__) // Not all build servers have a new enough glibc? TODO: remove when they're on gprecise.
    343   // Emulator kernels don't currently support setting the name of other threads.
    344   char* filename = NULL;
    345   asprintf(&filename, "/proc/self/task/%d/comm", gettid());
    346   struct stat sb;
    347   bool has_comm = (stat(filename, &sb) != -1);
    348   free(filename);
    349 
    350   if (has_comm) {
    351     pthread_t t1;
    352     ASSERT_EQ(0, pthread_create(&t1, NULL, SleepFn, reinterpret_cast<void*>(5)));
    353     ASSERT_EQ(0, pthread_setname_np(t1, "short 2"));
    354   } else {
    355     fprintf(stderr, "skipping test: this kernel doesn't have /proc/self/task/tid/comm files!\n");
    356   }
    357 #else // __BIONIC__
    358   GTEST_LOG_(INFO) << "This test does nothing.\n";
    359 #endif // __BIONIC__
    360 }
    361 
    362 TEST(pthread, pthread_setname_np__no_such_thread) {
    363 #if defined(__BIONIC__) // Not all build servers have a new enough glibc? TODO: remove when they're on gprecise.
    364   pthread_t dead_thread;
    365   MakeDeadThread(dead_thread);
    366 
    367   // Call pthread_setname_np after thread has already exited.
    368   ASSERT_EQ(ESRCH, pthread_setname_np(dead_thread, "short 3"));
    369 #else // __BIONIC__
    370   GTEST_LOG_(INFO) << "This test does nothing.\n";
    371 #endif // __BIONIC__
    372 }
    373 
    374 TEST(pthread, pthread_kill__0) {
    375   // Signal 0 just tests that the thread exists, so it's safe to call on ourselves.
    376   ASSERT_EQ(0, pthread_kill(pthread_self(), 0));
    377 }
    378 
    379 TEST(pthread, pthread_kill__invalid_signal) {
    380   ASSERT_EQ(EINVAL, pthread_kill(pthread_self(), -1));
    381 }
    382 
    383 static void pthread_kill__in_signal_handler_helper(int signal_number) {
    384   static int count = 0;
    385   ASSERT_EQ(SIGALRM, signal_number);
    386   if (++count == 1) {
    387     // Can we call pthread_kill from a signal handler?
    388     ASSERT_EQ(0, pthread_kill(pthread_self(), SIGALRM));
    389   }
    390 }
    391 
    392 TEST(pthread, pthread_kill__in_signal_handler) {
    393   ScopedSignalHandler ssh(SIGALRM, pthread_kill__in_signal_handler_helper);
    394   ASSERT_EQ(0, pthread_kill(pthread_self(), SIGALRM));
    395 }
    396 
    397 TEST(pthread, pthread_detach__no_such_thread) {
    398   pthread_t dead_thread;
    399   MakeDeadThread(dead_thread);
    400 
    401   ASSERT_EQ(ESRCH, pthread_detach(dead_thread));
    402 }
    403 
    404 TEST(pthread, pthread_detach__leak) {
    405   size_t initial_bytes = 0;
    406   // Run this loop more than once since the first loop causes some memory
    407   // to be allocated permenantly. Run an extra loop to help catch any subtle
    408   // memory leaks.
    409   for (size_t loop = 0; loop < 3; loop++) {
    410     // Set the initial bytes on the second loop since the memory in use
    411     // should have stabilized.
    412     if (loop == 1) {
    413       initial_bytes = mallinfo().uordblks;
    414     }
    415 
    416     pthread_attr_t attr;
    417     ASSERT_EQ(0, pthread_attr_init(&attr));
    418     ASSERT_EQ(0, pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_JOINABLE));
    419 
    420     std::vector<pthread_t> threads;
    421     for (size_t i = 0; i < 32; ++i) {
    422       pthread_t t;
    423       ASSERT_EQ(0, pthread_create(&t, &attr, IdFn, NULL));
    424       threads.push_back(t);
    425     }
    426 
    427     sleep(1);
    428 
    429     for (size_t i = 0; i < 32; ++i) {
    430       ASSERT_EQ(0, pthread_detach(threads[i])) << i;
    431     }
    432   }
    433 
    434   size_t final_bytes = mallinfo().uordblks;
    435   int leaked_bytes = (final_bytes - initial_bytes);
    436 
    437   // User code (like this test) doesn't know how large pthread_internal_t is.
    438   // We can be pretty sure it's more than 128 bytes.
    439   ASSERT_LT(leaked_bytes, 32 /*threads*/ * 128 /*bytes*/);
    440 }
    441 
    442 TEST(pthread, pthread_getcpuclockid__clock_gettime) {
    443   pthread_t t;
    444   ASSERT_EQ(0, pthread_create(&t, NULL, SleepFn, reinterpret_cast<void*>(5)));
    445 
    446   clockid_t c;
    447   ASSERT_EQ(0, pthread_getcpuclockid(t, &c));
    448   timespec ts;
    449   ASSERT_EQ(0, clock_gettime(c, &ts));
    450 }
    451 
    452 TEST(pthread, pthread_getcpuclockid__no_such_thread) {
    453   pthread_t dead_thread;
    454   MakeDeadThread(dead_thread);
    455 
    456   clockid_t c;
    457   ASSERT_EQ(ESRCH, pthread_getcpuclockid(dead_thread, &c));
    458 }
    459 
    460 TEST(pthread, pthread_getschedparam__no_such_thread) {
    461   pthread_t dead_thread;
    462   MakeDeadThread(dead_thread);
    463 
    464   int policy;
    465   sched_param param;
    466   ASSERT_EQ(ESRCH, pthread_getschedparam(dead_thread, &policy, &param));
    467 }
    468 
    469 TEST(pthread, pthread_setschedparam__no_such_thread) {
    470   pthread_t dead_thread;
    471   MakeDeadThread(dead_thread);
    472 
    473   int policy = 0;
    474   sched_param param;
    475   ASSERT_EQ(ESRCH, pthread_setschedparam(dead_thread, policy, &param));
    476 }
    477 
    478 TEST(pthread, pthread_join__no_such_thread) {
    479   pthread_t dead_thread;
    480   MakeDeadThread(dead_thread);
    481 
    482   void* result;
    483   ASSERT_EQ(ESRCH, pthread_join(dead_thread, &result));
    484 }
    485 
    486 TEST(pthread, pthread_kill__no_such_thread) {
    487   pthread_t dead_thread;
    488   MakeDeadThread(dead_thread);
    489 
    490   ASSERT_EQ(ESRCH, pthread_kill(dead_thread, 0));
    491 }
    492 
    493 TEST(pthread, pthread_join__multijoin) {
    494   bool done = false;
    495 
    496   pthread_t t1;
    497   ASSERT_EQ(0, pthread_create(&t1, NULL, SpinFn, &done));
    498 
    499   pthread_t t2;
    500   ASSERT_EQ(0, pthread_create(&t2, NULL, JoinFn, reinterpret_cast<void*>(t1)));
    501 
    502   sleep(1); // (Give t2 a chance to call pthread_join.)
    503 
    504   // Multiple joins to the same thread should fail.
    505   ASSERT_EQ(EINVAL, pthread_join(t1, NULL));
    506 
    507   done = true;
    508 
    509   // ...but t2's join on t1 still goes ahead (which we can tell because our join on t2 finishes).
    510   void* join_result;
    511   ASSERT_EQ(0, pthread_join(t2, &join_result));
    512   ASSERT_EQ(0U, reinterpret_cast<uintptr_t>(join_result));
    513 }
    514 
    515 TEST(pthread, pthread_join__race) {
    516   // http://b/11693195 --- pthread_join could return before the thread had actually exited.
    517   // If the joiner unmapped the thread's stack, that could lead to SIGSEGV in the thread.
    518   for (size_t i = 0; i < 1024; ++i) {
    519     size_t stack_size = 64*1024;
    520     void* stack = mmap(NULL, stack_size, PROT_READ|PROT_WRITE, MAP_ANON|MAP_PRIVATE, -1, 0);
    521 
    522     pthread_attr_t a;
    523     pthread_attr_init(&a);
    524     pthread_attr_setstack(&a, stack, stack_size);
    525 
    526     pthread_t t;
    527     ASSERT_EQ(0, pthread_create(&t, &a, IdFn, NULL));
    528     ASSERT_EQ(0, pthread_join(t, NULL));
    529     ASSERT_EQ(0, munmap(stack, stack_size));
    530   }
    531 }
    532 
    533 static void* GetActualGuardSizeFn(void* arg) {
    534   pthread_attr_t attributes;
    535   pthread_getattr_np(pthread_self(), &attributes);
    536   pthread_attr_getguardsize(&attributes, reinterpret_cast<size_t*>(arg));
    537   return NULL;
    538 }
    539 
    540 static size_t GetActualGuardSize(const pthread_attr_t& attributes) {
    541   size_t result;
    542   pthread_t t;
    543   pthread_create(&t, &attributes, GetActualGuardSizeFn, &result);
    544   void* join_result;
    545   pthread_join(t, &join_result);
    546   return result;
    547 }
    548 
    549 static void* GetActualStackSizeFn(void* arg) {
    550   pthread_attr_t attributes;
    551   pthread_getattr_np(pthread_self(), &attributes);
    552   pthread_attr_getstacksize(&attributes, reinterpret_cast<size_t*>(arg));
    553   return NULL;
    554 }
    555 
    556 static size_t GetActualStackSize(const pthread_attr_t& attributes) {
    557   size_t result;
    558   pthread_t t;
    559   pthread_create(&t, &attributes, GetActualStackSizeFn, &result);
    560   void* join_result;
    561   pthread_join(t, &join_result);
    562   return result;
    563 }
    564 
    565 TEST(pthread, pthread_attr_setguardsize) {
    566   pthread_attr_t attributes;
    567   ASSERT_EQ(0, pthread_attr_init(&attributes));
    568 
    569   // Get the default guard size.
    570   size_t default_guard_size;
    571   ASSERT_EQ(0, pthread_attr_getguardsize(&attributes, &default_guard_size));
    572 
    573   // No such thing as too small: will be rounded up to one page by pthread_create.
    574   ASSERT_EQ(0, pthread_attr_setguardsize(&attributes, 128));
    575   size_t guard_size;
    576   ASSERT_EQ(0, pthread_attr_getguardsize(&attributes, &guard_size));
    577   ASSERT_EQ(128U, guard_size);
    578   ASSERT_EQ(4096U, GetActualGuardSize(attributes));
    579 
    580   // Large enough and a multiple of the page size.
    581   ASSERT_EQ(0, pthread_attr_setguardsize(&attributes, 32*1024));
    582   ASSERT_EQ(0, pthread_attr_getguardsize(&attributes, &guard_size));
    583   ASSERT_EQ(32*1024U, guard_size);
    584 
    585   // Large enough but not a multiple of the page size; will be rounded up by pthread_create.
    586   ASSERT_EQ(0, pthread_attr_setguardsize(&attributes, 32*1024 + 1));
    587   ASSERT_EQ(0, pthread_attr_getguardsize(&attributes, &guard_size));
    588   ASSERT_EQ(32*1024U + 1, guard_size);
    589 }
    590 
    591 TEST(pthread, pthread_attr_setstacksize) {
    592   pthread_attr_t attributes;
    593   ASSERT_EQ(0, pthread_attr_init(&attributes));
    594 
    595   // Get the default stack size.
    596   size_t default_stack_size;
    597   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &default_stack_size));
    598 
    599   // Too small.
    600   ASSERT_EQ(EINVAL, pthread_attr_setstacksize(&attributes, 128));
    601   size_t stack_size;
    602   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &stack_size));
    603   ASSERT_EQ(default_stack_size, stack_size);
    604   ASSERT_GE(GetActualStackSize(attributes), default_stack_size);
    605 
    606   // Large enough and a multiple of the page size.
    607   ASSERT_EQ(0, pthread_attr_setstacksize(&attributes, 32*1024));
    608   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &stack_size));
    609   ASSERT_EQ(32*1024U, stack_size);
    610   ASSERT_EQ(GetActualStackSize(attributes), 32*1024U);
    611 
    612   // Large enough but not a multiple of the page size; will be rounded up by pthread_create.
    613   ASSERT_EQ(0, pthread_attr_setstacksize(&attributes, 32*1024 + 1));
    614   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &stack_size));
    615   ASSERT_EQ(32*1024U + 1, stack_size);
    616 #if defined(__BIONIC__)
    617   // Bionic rounds up, which is what POSIX allows.
    618   ASSERT_EQ(GetActualStackSize(attributes), (32 + 4)*1024U);
    619 #else // __BIONIC__
    620   // glibc rounds down, in violation of POSIX. They document this in their BUGS section.
    621   ASSERT_EQ(GetActualStackSize(attributes), 32*1024U);
    622 #endif // __BIONIC__
    623 }
    624 
    625 TEST(pthread, pthread_rwlock_smoke) {
    626   pthread_rwlock_t l;
    627   ASSERT_EQ(0, pthread_rwlock_init(&l, NULL));
    628 
    629   // Single read lock
    630   ASSERT_EQ(0, pthread_rwlock_rdlock(&l));
    631   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    632 
    633   // Multiple read lock
    634   ASSERT_EQ(0, pthread_rwlock_rdlock(&l));
    635   ASSERT_EQ(0, pthread_rwlock_rdlock(&l));
    636   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    637   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    638 
    639   // Write lock
    640   ASSERT_EQ(0, pthread_rwlock_wrlock(&l));
    641   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    642 
    643   // Try writer lock
    644   ASSERT_EQ(0, pthread_rwlock_trywrlock(&l));
    645   ASSERT_EQ(EBUSY, pthread_rwlock_trywrlock(&l));
    646   ASSERT_EQ(EBUSY, pthread_rwlock_tryrdlock(&l));
    647   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    648 
    649   // Try reader lock
    650   ASSERT_EQ(0, pthread_rwlock_tryrdlock(&l));
    651   ASSERT_EQ(0, pthread_rwlock_tryrdlock(&l));
    652   ASSERT_EQ(EBUSY, pthread_rwlock_trywrlock(&l));
    653   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    654   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    655 
    656   // Try writer lock after unlock
    657   ASSERT_EQ(0, pthread_rwlock_wrlock(&l));
    658   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    659 
    660 #ifdef __BIONIC__
    661   // EDEADLK in "read after write"
    662   ASSERT_EQ(0, pthread_rwlock_wrlock(&l));
    663   ASSERT_EQ(EDEADLK, pthread_rwlock_rdlock(&l));
    664   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    665 
    666   // EDEADLK in "write after write"
    667   ASSERT_EQ(0, pthread_rwlock_wrlock(&l));
    668   ASSERT_EQ(EDEADLK, pthread_rwlock_wrlock(&l));
    669   ASSERT_EQ(0, pthread_rwlock_unlock(&l));
    670 #endif
    671 
    672   ASSERT_EQ(0, pthread_rwlock_destroy(&l));
    673 }
    674 
    675 static int g_once_fn_call_count = 0;
    676 static void OnceFn() {
    677   ++g_once_fn_call_count;
    678 }
    679 
    680 TEST(pthread, pthread_once_smoke) {
    681   pthread_once_t once_control = PTHREAD_ONCE_INIT;
    682   ASSERT_EQ(0, pthread_once(&once_control, OnceFn));
    683   ASSERT_EQ(0, pthread_once(&once_control, OnceFn));
    684   ASSERT_EQ(1, g_once_fn_call_count);
    685 }
    686 
    687 static std::string pthread_once_1934122_result = "";
    688 
    689 static void Routine2() {
    690   pthread_once_1934122_result += "2";
    691 }
    692 
    693 static void Routine1() {
    694   pthread_once_t once_control_2 = PTHREAD_ONCE_INIT;
    695   pthread_once_1934122_result += "1";
    696   pthread_once(&once_control_2, &Routine2);
    697 }
    698 
    699 TEST(pthread, pthread_once_1934122) {
    700   // Very old versions of Android couldn't call pthread_once from a
    701   // pthread_once init routine. http://b/1934122.
    702   pthread_once_t once_control_1 = PTHREAD_ONCE_INIT;
    703   ASSERT_EQ(0, pthread_once(&once_control_1, &Routine1));
    704   ASSERT_EQ("12", pthread_once_1934122_result);
    705 }
    706 
    707 static int g_atfork_prepare_calls = 0;
    708 static void AtForkPrepare1() { g_atfork_prepare_calls = (g_atfork_prepare_calls << 4) | 1; }
    709 static void AtForkPrepare2() { g_atfork_prepare_calls = (g_atfork_prepare_calls << 4) | 2; }
    710 static int g_atfork_parent_calls = 0;
    711 static void AtForkParent1() { g_atfork_parent_calls = (g_atfork_parent_calls << 4) | 1; }
    712 static void AtForkParent2() { g_atfork_parent_calls = (g_atfork_parent_calls << 4) | 2; }
    713 static int g_atfork_child_calls = 0;
    714 static void AtForkChild1() { g_atfork_child_calls = (g_atfork_child_calls << 4) | 1; }
    715 static void AtForkChild2() { g_atfork_child_calls = (g_atfork_child_calls << 4) | 2; }
    716 
    717 TEST(pthread, pthread_atfork) {
    718   ASSERT_EQ(0, pthread_atfork(AtForkPrepare1, AtForkParent1, AtForkChild1));
    719   ASSERT_EQ(0, pthread_atfork(AtForkPrepare2, AtForkParent2, AtForkChild2));
    720 
    721   int pid = fork();
    722   ASSERT_NE(-1, pid) << strerror(errno);
    723 
    724   // Child and parent calls are made in the order they were registered.
    725   if (pid == 0) {
    726     ASSERT_EQ(0x12, g_atfork_child_calls);
    727     _exit(0);
    728   }
    729   ASSERT_EQ(0x12, g_atfork_parent_calls);
    730 
    731   // Prepare calls are made in the reverse order.
    732   ASSERT_EQ(0x21, g_atfork_prepare_calls);
    733 }
    734 
    735 TEST(pthread, pthread_attr_getscope) {
    736   pthread_attr_t attr;
    737   ASSERT_EQ(0, pthread_attr_init(&attr));
    738 
    739   int scope;
    740   ASSERT_EQ(0, pthread_attr_getscope(&attr, &scope));
    741   ASSERT_EQ(PTHREAD_SCOPE_SYSTEM, scope);
    742 }
    743 
    744 TEST(pthread, pthread_condattr_init) {
    745   pthread_condattr_t attr;
    746   pthread_condattr_init(&attr);
    747 
    748   clockid_t clock;
    749   ASSERT_EQ(0, pthread_condattr_getclock(&attr, &clock));
    750   ASSERT_EQ(CLOCK_REALTIME, clock);
    751 
    752   int pshared;
    753   ASSERT_EQ(0, pthread_condattr_getpshared(&attr, &pshared));
    754   ASSERT_EQ(PTHREAD_PROCESS_PRIVATE, pshared);
    755 }
    756 
    757 TEST(pthread, pthread_condattr_setclock) {
    758   pthread_condattr_t attr;
    759   pthread_condattr_init(&attr);
    760 
    761   ASSERT_EQ(0, pthread_condattr_setclock(&attr, CLOCK_REALTIME));
    762   clockid_t clock;
    763   ASSERT_EQ(0, pthread_condattr_getclock(&attr, &clock));
    764   ASSERT_EQ(CLOCK_REALTIME, clock);
    765 
    766   ASSERT_EQ(0, pthread_condattr_setclock(&attr, CLOCK_MONOTONIC));
    767   ASSERT_EQ(0, pthread_condattr_getclock(&attr, &clock));
    768   ASSERT_EQ(CLOCK_MONOTONIC, clock);
    769 
    770   ASSERT_EQ(EINVAL, pthread_condattr_setclock(&attr, CLOCK_PROCESS_CPUTIME_ID));
    771 }
    772 
    773 TEST(pthread, pthread_cond_broadcast__preserves_condattr_flags) {
    774 #if defined(__BIONIC__) // This tests a bionic implementation detail.
    775   pthread_condattr_t attr;
    776   pthread_condattr_init(&attr);
    777 
    778   ASSERT_EQ(0, pthread_condattr_setclock(&attr, CLOCK_MONOTONIC));
    779   ASSERT_EQ(0, pthread_condattr_setpshared(&attr, PTHREAD_PROCESS_SHARED));
    780 
    781   pthread_cond_t cond_var;
    782   ASSERT_EQ(0, pthread_cond_init(&cond_var, &attr));
    783 
    784   ASSERT_EQ(0, pthread_cond_signal(&cond_var));
    785   ASSERT_EQ(0, pthread_cond_broadcast(&cond_var));
    786 
    787   attr = static_cast<pthread_condattr_t>(cond_var.value);
    788   clockid_t clock;
    789   ASSERT_EQ(0, pthread_condattr_getclock(&attr, &clock));
    790   ASSERT_EQ(CLOCK_MONOTONIC, clock);
    791   int pshared;
    792   ASSERT_EQ(0, pthread_condattr_getpshared(&attr, &pshared));
    793   ASSERT_EQ(PTHREAD_PROCESS_SHARED, pshared);
    794 #else // __BIONIC__
    795   GTEST_LOG_(INFO) << "This test does nothing.\n";
    796 #endif // __BIONIC__
    797 }
    798 
    799 TEST(pthread, pthread_mutex_timedlock) {
    800   pthread_mutex_t m;
    801   ASSERT_EQ(0, pthread_mutex_init(&m, NULL));
    802 
    803   // If the mutex is already locked, pthread_mutex_timedlock should time out.
    804   ASSERT_EQ(0, pthread_mutex_lock(&m));
    805 
    806   timespec ts;
    807   ASSERT_EQ(0, clock_gettime(CLOCK_REALTIME, &ts));
    808   ts.tv_nsec += 1;
    809   ASSERT_EQ(ETIMEDOUT, pthread_mutex_timedlock(&m, &ts));
    810 
    811   // If the mutex is unlocked, pthread_mutex_timedlock should succeed.
    812   ASSERT_EQ(0, pthread_mutex_unlock(&m));
    813 
    814   ASSERT_EQ(0, clock_gettime(CLOCK_REALTIME, &ts));
    815   ts.tv_nsec += 1;
    816   ASSERT_EQ(0, pthread_mutex_timedlock(&m, &ts));
    817 
    818   ASSERT_EQ(0, pthread_mutex_unlock(&m));
    819   ASSERT_EQ(0, pthread_mutex_destroy(&m));
    820 }
    821 
    822 TEST(pthread, pthread_attr_getstack__main_thread) {
    823   // This test is only meaningful for the main thread, so make sure we're running on it!
    824   ASSERT_EQ(getpid(), syscall(__NR_gettid));
    825 
    826   // Get the main thread's attributes.
    827   pthread_attr_t attributes;
    828   ASSERT_EQ(0, pthread_getattr_np(pthread_self(), &attributes));
    829 
    830   // Check that we correctly report that the main thread has no guard page.
    831   size_t guard_size;
    832   ASSERT_EQ(0, pthread_attr_getguardsize(&attributes, &guard_size));
    833   ASSERT_EQ(0U, guard_size); // The main thread has no guard page.
    834 
    835   // Get the stack base and the stack size (both ways).
    836   void* stack_base;
    837   size_t stack_size;
    838   ASSERT_EQ(0, pthread_attr_getstack(&attributes, &stack_base, &stack_size));
    839   size_t stack_size2;
    840   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &stack_size2));
    841 
    842   // The two methods of asking for the stack size should agree.
    843   EXPECT_EQ(stack_size, stack_size2);
    844 
    845   // What does /proc/self/maps' [stack] line say?
    846   void* maps_stack_hi = NULL;
    847   FILE* fp = fopen("/proc/self/maps", "r");
    848   ASSERT_TRUE(fp != NULL);
    849   char line[BUFSIZ];
    850   while (fgets(line, sizeof(line), fp) != NULL) {
    851     uintptr_t lo, hi;
    852     char name[10];
    853     sscanf(line, "%" PRIxPTR "-%" PRIxPTR " %*4s %*x %*x:%*x %*d %10s", &lo, &hi, name);
    854     if (strcmp(name, "[stack]") == 0) {
    855       maps_stack_hi = reinterpret_cast<void*>(hi);
    856       break;
    857     }
    858   }
    859   fclose(fp);
    860 
    861   // The stack size should correspond to RLIMIT_STACK.
    862   rlimit rl;
    863   ASSERT_EQ(0, getrlimit(RLIMIT_STACK, &rl));
    864   uint64_t original_rlim_cur = rl.rlim_cur;
    865 #if defined(__BIONIC__)
    866   if (rl.rlim_cur == RLIM_INFINITY) {
    867     rl.rlim_cur = 8 * 1024 * 1024; // Bionic reports unlimited stacks as 8MiB.
    868   }
    869 #endif
    870   EXPECT_EQ(rl.rlim_cur, stack_size);
    871 
    872   auto guard = create_scope_guard([&rl, original_rlim_cur]() {
    873     rl.rlim_cur = original_rlim_cur;
    874     ASSERT_EQ(0, setrlimit(RLIMIT_STACK, &rl));
    875   });
    876 
    877   // The high address of the /proc/self/maps [stack] region should equal stack_base + stack_size.
    878   // Remember that the stack grows down (and is mapped in on demand), so the low address of the
    879   // region isn't very interesting.
    880   EXPECT_EQ(maps_stack_hi, reinterpret_cast<uint8_t*>(stack_base) + stack_size);
    881 
    882   //
    883   // What if RLIMIT_STACK is smaller than the stack's current extent?
    884   //
    885   rl.rlim_cur = rl.rlim_max = 1024; // 1KiB. We know the stack must be at least a page already.
    886   rl.rlim_max = RLIM_INFINITY;
    887   ASSERT_EQ(0, setrlimit(RLIMIT_STACK, &rl));
    888 
    889   ASSERT_EQ(0, pthread_getattr_np(pthread_self(), &attributes));
    890   ASSERT_EQ(0, pthread_attr_getstack(&attributes, &stack_base, &stack_size));
    891   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &stack_size2));
    892 
    893   EXPECT_EQ(stack_size, stack_size2);
    894   ASSERT_EQ(1024U, stack_size);
    895 
    896   //
    897   // What if RLIMIT_STACK isn't a whole number of pages?
    898   //
    899   rl.rlim_cur = rl.rlim_max = 6666; // Not a whole number of pages.
    900   rl.rlim_max = RLIM_INFINITY;
    901   ASSERT_EQ(0, setrlimit(RLIMIT_STACK, &rl));
    902 
    903   ASSERT_EQ(0, pthread_getattr_np(pthread_self(), &attributes));
    904   ASSERT_EQ(0, pthread_attr_getstack(&attributes, &stack_base, &stack_size));
    905   ASSERT_EQ(0, pthread_attr_getstacksize(&attributes, &stack_size2));
    906 
    907   EXPECT_EQ(stack_size, stack_size2);
    908   ASSERT_EQ(6666U, stack_size);
    909 }
    910