Home | History | Annotate | Download | only in cctest
      1 // Copyright 2012 the V8 project authors. All rights reserved.
      2 // Redistribution and use in source and binary forms, with or without
      3 // modification, are permitted provided that the following conditions are
      4 // met:
      5 //
      6 //     * Redistributions of source code must retain the above copyright
      7 //       notice, this list of conditions and the following disclaimer.
      8 //     * Redistributions in binary form must reproduce the above
      9 //       copyright notice, this list of conditions and the following
     10 //       disclaimer in the documentation and/or other materials provided
     11 //       with the distribution.
     12 //     * Neither the name of Google Inc. nor the names of its
     13 //       contributors may be used to endorse or promote products derived
     14 //       from this software without specific prior written permission.
     15 //
     16 // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     17 // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     18 // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
     19 // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
     20 // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
     21 // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
     22 // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     23 // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
     26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     27 
     28 #include <stdlib.h>
     29 
     30 #ifdef __linux__
     31 #include <sys/types.h>
     32 #include <sys/stat.h>
     33 #include <fcntl.h>
     34 #include <unistd.h>
     35 #include <errno.h>
     36 #endif
     37 
     38 #include "v8.h"
     39 
     40 #include "global-handles.h"
     41 #include "snapshot.h"
     42 #include "cctest.h"
     43 
     44 using namespace v8::internal;
     45 
     46 static v8::Persistent<v8::Context> env;
     47 
     48 static void InitializeVM() {
     49   if (env.IsEmpty()) env = v8::Context::New();
     50   v8::HandleScope scope;
     51   env->Enter();
     52 }
     53 
     54 
     55 TEST(MarkingDeque) {
     56   int mem_size = 20 * kPointerSize;
     57   byte* mem = NewArray<byte>(20*kPointerSize);
     58   Address low = reinterpret_cast<Address>(mem);
     59   Address high = low + mem_size;
     60   MarkingDeque s;
     61   s.Initialize(low, high);
     62 
     63   Address address = NULL;
     64   while (!s.IsFull()) {
     65     s.PushBlack(HeapObject::FromAddress(address));
     66     address += kPointerSize;
     67   }
     68 
     69   while (!s.IsEmpty()) {
     70     Address value = s.Pop()->address();
     71     address -= kPointerSize;
     72     CHECK_EQ(address, value);
     73   }
     74 
     75   CHECK_EQ(NULL, address);
     76   DeleteArray(mem);
     77 }
     78 
     79 
     80 TEST(Promotion) {
     81   // This test requires compaction. If compaction is turned off, we
     82   // skip the entire test.
     83   if (FLAG_never_compact) return;
     84 
     85   // Ensure that we get a compacting collection so that objects are promoted
     86   // from new space.
     87   FLAG_gc_global = true;
     88   FLAG_always_compact = true;
     89   HEAP->ConfigureHeap(2*256*KB, 8*MB, 8*MB);
     90 
     91   InitializeVM();
     92 
     93   v8::HandleScope sc;
     94 
     95   // Allocate a fixed array in the new space.
     96   int array_size =
     97       (Page::kMaxNonCodeHeapObjectSize - FixedArray::kHeaderSize) /
     98       (kPointerSize * 4);
     99   Object* obj = HEAP->AllocateFixedArray(array_size)->ToObjectChecked();
    100 
    101   Handle<FixedArray> array(FixedArray::cast(obj));
    102 
    103   // Array should be in the new space.
    104   CHECK(HEAP->InSpace(*array, NEW_SPACE));
    105 
    106   // Call the m-c collector, so array becomes an old object.
    107   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    108 
    109   // Array now sits in the old space
    110   CHECK(HEAP->InSpace(*array, OLD_POINTER_SPACE));
    111 }
    112 
    113 
    114 TEST(NoPromotion) {
    115   HEAP->ConfigureHeap(2*256*KB, 8*MB, 8*MB);
    116 
    117   // Test the situation that some objects in new space are promoted to
    118   // the old space
    119   InitializeVM();
    120 
    121   v8::HandleScope sc;
    122 
    123   // Do a mark compact GC to shrink the heap.
    124   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    125 
    126   // Allocate a big Fixed array in the new space.
    127   int max_size =
    128       Min(Page::kMaxNonCodeHeapObjectSize, HEAP->MaxObjectSizeInNewSpace());
    129 
    130   int length = (max_size - FixedArray::kHeaderSize) / (2*kPointerSize);
    131   Object* obj = i::Isolate::Current()->heap()->AllocateFixedArray(length)->
    132       ToObjectChecked();
    133 
    134   Handle<FixedArray> array(FixedArray::cast(obj));
    135 
    136   // Array still stays in the new space.
    137   CHECK(HEAP->InSpace(*array, NEW_SPACE));
    138 
    139   // Allocate objects in the old space until out of memory.
    140   FixedArray* host = *array;
    141   while (true) {
    142     Object* obj;
    143     { MaybeObject* maybe_obj = HEAP->AllocateFixedArray(100, TENURED);
    144       if (!maybe_obj->ToObject(&obj)) break;
    145     }
    146 
    147     host->set(0, obj);
    148     host = FixedArray::cast(obj);
    149   }
    150 
    151   // Call mark compact GC, and it should pass.
    152   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    153 }
    154 
    155 
    156 TEST(MarkCompactCollector) {
    157   InitializeVM();
    158 
    159   v8::HandleScope sc;
    160   // call mark-compact when heap is empty
    161   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    162 
    163   // keep allocating garbage in new space until it fails
    164   const int ARRAY_SIZE = 100;
    165   Object* array;
    166   MaybeObject* maybe_array;
    167   do {
    168     maybe_array = HEAP->AllocateFixedArray(ARRAY_SIZE);
    169   } while (maybe_array->ToObject(&array));
    170   HEAP->CollectGarbage(NEW_SPACE);
    171 
    172   array = HEAP->AllocateFixedArray(ARRAY_SIZE)->ToObjectChecked();
    173 
    174   // keep allocating maps until it fails
    175   Object* mapp;
    176   MaybeObject* maybe_mapp;
    177   do {
    178     maybe_mapp = HEAP->AllocateMap(JS_OBJECT_TYPE, JSObject::kHeaderSize);
    179   } while (maybe_mapp->ToObject(&mapp));
    180   HEAP->CollectGarbage(MAP_SPACE);
    181   mapp = HEAP->AllocateMap(JS_OBJECT_TYPE,
    182                            JSObject::kHeaderSize)->ToObjectChecked();
    183 
    184   // allocate a garbage
    185   String* func_name =
    186       String::cast(HEAP->LookupAsciiSymbol("theFunction")->ToObjectChecked());
    187   SharedFunctionInfo* function_share = SharedFunctionInfo::cast(
    188       HEAP->AllocateSharedFunctionInfo(func_name)->ToObjectChecked());
    189   JSFunction* function = JSFunction::cast(
    190       HEAP->AllocateFunction(*Isolate::Current()->function_map(),
    191                              function_share,
    192                              HEAP->undefined_value())->ToObjectChecked());
    193   Map* initial_map =
    194       Map::cast(HEAP->AllocateMap(JS_OBJECT_TYPE,
    195                                   JSObject::kHeaderSize)->ToObjectChecked());
    196   function->set_initial_map(initial_map);
    197   Isolate::Current()->context()->global()->SetProperty(
    198       func_name, function, NONE, kNonStrictMode)->ToObjectChecked();
    199 
    200   JSObject* obj = JSObject::cast(
    201       HEAP->AllocateJSObject(function)->ToObjectChecked());
    202   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    203 
    204   func_name =
    205       String::cast(HEAP->LookupAsciiSymbol("theFunction")->ToObjectChecked());
    206   CHECK(Isolate::Current()->context()->global()->HasLocalProperty(func_name));
    207   Object* func_value = Isolate::Current()->context()->global()->
    208       GetProperty(func_name)->ToObjectChecked();
    209   CHECK(func_value->IsJSFunction());
    210   function = JSFunction::cast(func_value);
    211 
    212   obj = JSObject::cast(HEAP->AllocateJSObject(function)->ToObjectChecked());
    213   String* obj_name =
    214       String::cast(HEAP->LookupAsciiSymbol("theObject")->ToObjectChecked());
    215   Isolate::Current()->context()->global()->SetProperty(
    216       obj_name, obj, NONE, kNonStrictMode)->ToObjectChecked();
    217   String* prop_name =
    218       String::cast(HEAP->LookupAsciiSymbol("theSlot")->ToObjectChecked());
    219   obj->SetProperty(prop_name,
    220                    Smi::FromInt(23),
    221                    NONE,
    222                    kNonStrictMode)->ToObjectChecked();
    223 
    224   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    225 
    226   obj_name =
    227       String::cast(HEAP->LookupAsciiSymbol("theObject")->ToObjectChecked());
    228   CHECK(Isolate::Current()->context()->global()->HasLocalProperty(obj_name));
    229   CHECK(Isolate::Current()->context()->global()->
    230         GetProperty(obj_name)->ToObjectChecked()->IsJSObject());
    231   obj = JSObject::cast(Isolate::Current()->context()->global()->
    232                        GetProperty(obj_name)->ToObjectChecked());
    233   prop_name =
    234       String::cast(HEAP->LookupAsciiSymbol("theSlot")->ToObjectChecked());
    235   CHECK(obj->GetProperty(prop_name) == Smi::FromInt(23));
    236 }
    237 
    238 
    239 // TODO(1600): compaction of map space is temporary removed from GC.
    240 #if 0
    241 static Handle<Map> CreateMap() {
    242   return FACTORY->NewMap(JS_OBJECT_TYPE, JSObject::kHeaderSize);
    243 }
    244 
    245 
    246 TEST(MapCompact) {
    247   FLAG_max_map_space_pages = 16;
    248   InitializeVM();
    249 
    250   {
    251     v8::HandleScope sc;
    252     // keep allocating maps while pointers are still encodable and thus
    253     // mark compact is permitted.
    254     Handle<JSObject> root = FACTORY->NewJSObjectFromMap(CreateMap());
    255     do {
    256       Handle<Map> map = CreateMap();
    257       map->set_prototype(*root);
    258       root = FACTORY->NewJSObjectFromMap(map);
    259     } while (HEAP->map_space()->MapPointersEncodable());
    260   }
    261   // Now, as we don't have any handles to just allocated maps, we should
    262   // be able to trigger map compaction.
    263   // To give an additional chance to fail, try to force compaction which
    264   // should be impossible right now.
    265   HEAP->CollectAllGarbage(Heap::kForceCompactionMask);
    266   // And now map pointers should be encodable again.
    267   CHECK(HEAP->map_space()->MapPointersEncodable());
    268 }
    269 #endif
    270 
    271 static int gc_starts = 0;
    272 static int gc_ends = 0;
    273 
    274 static void GCPrologueCallbackFunc() {
    275   CHECK(gc_starts == gc_ends);
    276   gc_starts++;
    277 }
    278 
    279 
    280 static void GCEpilogueCallbackFunc() {
    281   CHECK(gc_starts == gc_ends + 1);
    282   gc_ends++;
    283 }
    284 
    285 
    286 TEST(GCCallback) {
    287   InitializeVM();
    288 
    289   HEAP->SetGlobalGCPrologueCallback(&GCPrologueCallbackFunc);
    290   HEAP->SetGlobalGCEpilogueCallback(&GCEpilogueCallbackFunc);
    291 
    292   // Scavenge does not call GC callback functions.
    293   HEAP->PerformScavenge();
    294 
    295   CHECK_EQ(0, gc_starts);
    296   CHECK_EQ(gc_ends, gc_starts);
    297 
    298   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    299   CHECK_EQ(1, gc_starts);
    300   CHECK_EQ(gc_ends, gc_starts);
    301 }
    302 
    303 
    304 static int NumberOfWeakCalls = 0;
    305 static void WeakPointerCallback(v8::Persistent<v8::Value> handle, void* id) {
    306   ASSERT(id == reinterpret_cast<void*>(1234));
    307   NumberOfWeakCalls++;
    308   handle.Dispose();
    309 }
    310 
    311 TEST(ObjectGroups) {
    312   InitializeVM();
    313   GlobalHandles* global_handles = Isolate::Current()->global_handles();
    314 
    315   NumberOfWeakCalls = 0;
    316   v8::HandleScope handle_scope;
    317 
    318   Handle<Object> g1s1 =
    319       global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    320   Handle<Object> g1s2 =
    321       global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    322   Handle<Object> g1c1 =
    323       global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    324   global_handles->MakeWeak(g1s1.location(),
    325                            reinterpret_cast<void*>(1234),
    326                            &WeakPointerCallback);
    327   global_handles->MakeWeak(g1s2.location(),
    328                            reinterpret_cast<void*>(1234),
    329                            &WeakPointerCallback);
    330   global_handles->MakeWeak(g1c1.location(),
    331                            reinterpret_cast<void*>(1234),
    332                            &WeakPointerCallback);
    333 
    334   Handle<Object> g2s1 =
    335       global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    336   Handle<Object> g2s2 =
    337     global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    338   Handle<Object> g2c1 =
    339     global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    340   global_handles->MakeWeak(g2s1.location(),
    341                            reinterpret_cast<void*>(1234),
    342                            &WeakPointerCallback);
    343   global_handles->MakeWeak(g2s2.location(),
    344                            reinterpret_cast<void*>(1234),
    345                            &WeakPointerCallback);
    346   global_handles->MakeWeak(g2c1.location(),
    347                            reinterpret_cast<void*>(1234),
    348                            &WeakPointerCallback);
    349 
    350   Handle<Object> root = global_handles->Create(*g1s1);  // make a root.
    351 
    352   // Connect group 1 and 2, make a cycle.
    353   Handle<FixedArray>::cast(g1s2)->set(0, *g2s2);
    354   Handle<FixedArray>::cast(g2s1)->set(0, *g1s1);
    355 
    356   {
    357     Object** g1_objects[] = { g1s1.location(), g1s2.location() };
    358     Object** g1_children[] = { g1c1.location() };
    359     Object** g2_objects[] = { g2s1.location(), g2s2.location() };
    360     Object** g2_children[] = { g2c1.location() };
    361     global_handles->AddObjectGroup(g1_objects, 2, NULL);
    362     global_handles->AddImplicitReferences(
    363         Handle<HeapObject>::cast(g1s1).location(), g1_children, 1);
    364     global_handles->AddObjectGroup(g2_objects, 2, NULL);
    365     global_handles->AddImplicitReferences(
    366         Handle<HeapObject>::cast(g2s2).location(), g2_children, 1);
    367   }
    368   // Do a full GC
    369   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    370 
    371   // All object should be alive.
    372   CHECK_EQ(0, NumberOfWeakCalls);
    373 
    374   // Weaken the root.
    375   global_handles->MakeWeak(root.location(),
    376                            reinterpret_cast<void*>(1234),
    377                            &WeakPointerCallback);
    378   // But make children strong roots---all the objects (except for children)
    379   // should be collectable now.
    380   global_handles->ClearWeakness(g1c1.location());
    381   global_handles->ClearWeakness(g2c1.location());
    382 
    383   // Groups are deleted, rebuild groups.
    384   {
    385     Object** g1_objects[] = { g1s1.location(), g1s2.location() };
    386     Object** g1_children[] = { g1c1.location() };
    387     Object** g2_objects[] = { g2s1.location(), g2s2.location() };
    388     Object** g2_children[] = { g2c1.location() };
    389     global_handles->AddObjectGroup(g1_objects, 2, NULL);
    390     global_handles->AddImplicitReferences(
    391         Handle<HeapObject>::cast(g1s1).location(), g1_children, 1);
    392     global_handles->AddObjectGroup(g2_objects, 2, NULL);
    393     global_handles->AddImplicitReferences(
    394         Handle<HeapObject>::cast(g2s2).location(), g2_children, 1);
    395   }
    396 
    397   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    398 
    399   // All objects should be gone. 5 global handles in total.
    400   CHECK_EQ(5, NumberOfWeakCalls);
    401 
    402   // And now make children weak again and collect them.
    403   global_handles->MakeWeak(g1c1.location(),
    404                            reinterpret_cast<void*>(1234),
    405                            &WeakPointerCallback);
    406   global_handles->MakeWeak(g2c1.location(),
    407                            reinterpret_cast<void*>(1234),
    408                            &WeakPointerCallback);
    409 
    410   HEAP->CollectGarbage(OLD_POINTER_SPACE);
    411   CHECK_EQ(7, NumberOfWeakCalls);
    412 }
    413 
    414 
    415 class TestRetainedObjectInfo : public v8::RetainedObjectInfo {
    416  public:
    417   TestRetainedObjectInfo() : has_been_disposed_(false) {}
    418 
    419   bool has_been_disposed() { return has_been_disposed_; }
    420 
    421   virtual void Dispose() {
    422     ASSERT(!has_been_disposed_);
    423     has_been_disposed_ = true;
    424   }
    425 
    426   virtual bool IsEquivalent(v8::RetainedObjectInfo* other) {
    427     return other == this;
    428   }
    429 
    430   virtual intptr_t GetHash() { return 0; }
    431 
    432   virtual const char* GetLabel() { return "whatever"; }
    433 
    434  private:
    435   bool has_been_disposed_;
    436 };
    437 
    438 
    439 TEST(EmptyObjectGroups) {
    440   InitializeVM();
    441   GlobalHandles* global_handles = Isolate::Current()->global_handles();
    442 
    443   v8::HandleScope handle_scope;
    444 
    445   Handle<Object> object =
    446       global_handles->Create(HEAP->AllocateFixedArray(1)->ToObjectChecked());
    447 
    448   TestRetainedObjectInfo info;
    449   global_handles->AddObjectGroup(NULL, 0, &info);
    450   ASSERT(info.has_been_disposed());
    451 
    452   global_handles->AddImplicitReferences(
    453         Handle<HeapObject>::cast(object).location(), NULL, 0);
    454 }
    455 
    456 
    457 // Here is a memory use test that uses /proc, and is therefore Linux-only.  We
    458 // do not care how much memory the simulator uses, since it is only there for
    459 // debugging purposes.
    460 #if defined(__linux__) && !defined(USE_SIMULATOR)
    461 
    462 
    463 static uintptr_t ReadLong(char* buffer, intptr_t* position, int base) {
    464   char* end_address = buffer + *position;
    465   uintptr_t result = strtoul(buffer + *position, &end_address, base);
    466   CHECK(result != ULONG_MAX || errno != ERANGE);
    467   CHECK(end_address > buffer + *position);
    468   *position = end_address - buffer;
    469   return result;
    470 }
    471 
    472 
    473 static intptr_t MemoryInUse() {
    474   intptr_t memory_use = 0;
    475 
    476   int fd = open("/proc/self/maps", O_RDONLY);
    477   if (fd < 0) return -1;
    478 
    479   const int kBufSize = 10000;
    480   char buffer[kBufSize];
    481   int length = read(fd, buffer, kBufSize);
    482   intptr_t line_start = 0;
    483   CHECK_LT(length, kBufSize);  // Make the buffer bigger.
    484   CHECK_GT(length, 0);  // We have to find some data in the file.
    485   while (line_start < length) {
    486     if (buffer[line_start] == '\n') {
    487       line_start++;
    488       continue;
    489     }
    490     intptr_t position = line_start;
    491     uintptr_t start = ReadLong(buffer, &position, 16);
    492     CHECK_EQ(buffer[position++], '-');
    493     uintptr_t end = ReadLong(buffer, &position, 16);
    494     CHECK_EQ(buffer[position++], ' ');
    495     CHECK(buffer[position] == '-' || buffer[position] == 'r');
    496     bool read_permission = (buffer[position++] == 'r');
    497     CHECK(buffer[position] == '-' || buffer[position] == 'w');
    498     bool write_permission = (buffer[position++] == 'w');
    499     CHECK(buffer[position] == '-' || buffer[position] == 'x');
    500     bool execute_permission = (buffer[position++] == 'x');
    501     CHECK(buffer[position] == '-' || buffer[position] == 'p');
    502     bool private_mapping = (buffer[position++] == 'p');
    503     CHECK_EQ(buffer[position++], ' ');
    504     uintptr_t offset = ReadLong(buffer, &position, 16);
    505     USE(offset);
    506     CHECK_EQ(buffer[position++], ' ');
    507     uintptr_t major = ReadLong(buffer, &position, 16);
    508     USE(major);
    509     CHECK_EQ(buffer[position++], ':');
    510     uintptr_t minor = ReadLong(buffer, &position, 16);
    511     USE(minor);
    512     CHECK_EQ(buffer[position++], ' ');
    513     uintptr_t inode = ReadLong(buffer, &position, 10);
    514     while (position < length && buffer[position] != '\n') position++;
    515     if ((read_permission || write_permission || execute_permission) &&
    516         private_mapping && inode == 0) {
    517       memory_use += (end - start);
    518     }
    519 
    520     line_start = position;
    521   }
    522   close(fd);
    523   return memory_use;
    524 }
    525 
    526 
    527 TEST(BootUpMemoryUse) {
    528   intptr_t initial_memory = MemoryInUse();
    529   FLAG_crankshaft = false;  // Avoid flakiness.
    530   // Only Linux has the proc filesystem and only if it is mapped.  If it's not
    531   // there we just skip the test.
    532   if (initial_memory >= 0) {
    533     InitializeVM();
    534     intptr_t booted_memory = MemoryInUse();
    535     if (sizeof(initial_memory) == 8) {
    536       if (v8::internal::Snapshot::IsEnabled()) {
    537         CHECK_LE(booted_memory - initial_memory, 6686 * 1024);  // 6476.
    538       } else {
    539         CHECK_LE(booted_memory - initial_memory, 6809 * 1024);  // 6628.
    540       }
    541     } else {
    542       if (v8::internal::Snapshot::IsEnabled()) {
    543         CHECK_LE(booted_memory - initial_memory, 6532 * 1024);  // 6388.
    544       } else {
    545         CHECK_LE(booted_memory - initial_memory, 6940 * 1024);  // 6456
    546       }
    547     }
    548   }
    549 }
    550 
    551 #endif  // __linux__ and !USE_SIMULATOR
    552