1 // Copyright 2015 The Chromium Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style license that can be 3 // found in the LICENSE file. 4 5 #include "base/metrics/persistent_memory_allocator.h" 6 7 #include <memory> 8 9 #include "base/files/file.h" 10 #include "base/files/file_util.h" 11 #include "base/files/memory_mapped_file.h" 12 #include "base/files/scoped_temp_dir.h" 13 #include "base/memory/shared_memory.h" 14 #include "base/metrics/histogram.h" 15 #include "base/rand_util.h" 16 #include "base/strings/safe_sprintf.h" 17 #include "base/synchronization/condition_variable.h" 18 #include "base/synchronization/lock.h" 19 #include "base/threading/simple_thread.h" 20 #include "testing/gmock/include/gmock/gmock.h" 21 22 namespace { 23 24 const uint32_t TEST_MEMORY_SIZE = 1 << 20; // 1 MiB 25 const uint32_t TEST_MEMORY_PAGE = 64 << 10; // 64 KiB 26 const uint32_t TEST_ID = 12345; 27 const char TEST_NAME[] = "TestAllocator"; 28 29 } // namespace 30 31 namespace base { 32 33 typedef PersistentMemoryAllocator::Reference Reference; 34 35 class PersistentMemoryAllocatorTest : public testing::Test { 36 public: 37 // This can't be statically initialized because it's value isn't defined 38 // in the PersistentMemoryAllocator header file. Instead, it's simply set 39 // in the constructor. 40 uint32_t kAllocAlignment; 41 42 struct TestObject1 { 43 static constexpr uint32_t kPersistentTypeId = 1; 44 static constexpr size_t kExpectedInstanceSize = 4 + 1 + 3; 45 int32_t onething; 46 char oranother; 47 }; 48 49 struct TestObject2 { 50 static constexpr uint32_t kPersistentTypeId = 2; 51 static constexpr size_t kExpectedInstanceSize = 8 + 4 + 4 + 8 + 8; 52 int64_t thiis; 53 int32_t that; 54 float andthe; 55 double other; 56 char thing[8]; 57 }; 58 59 PersistentMemoryAllocatorTest() { 60 kAllocAlignment = GetAllocAlignment(); 61 mem_segment_.reset(new char[TEST_MEMORY_SIZE]); 62 } 63 64 void SetUp() override { 65 allocator_.reset(); 66 ::memset(mem_segment_.get(), 0, TEST_MEMORY_SIZE); 67 allocator_.reset(new PersistentMemoryAllocator( 68 mem_segment_.get(), TEST_MEMORY_SIZE, TEST_MEMORY_PAGE, 69 TEST_ID, TEST_NAME, false)); 70 } 71 72 void TearDown() override { 73 allocator_.reset(); 74 } 75 76 unsigned CountIterables() { 77 PersistentMemoryAllocator::Iterator iter(allocator_.get()); 78 uint32_t type; 79 unsigned count = 0; 80 while (iter.GetNext(&type) != 0) { 81 ++count; 82 } 83 return count; 84 } 85 86 static uint32_t GetAllocAlignment() { 87 return PersistentMemoryAllocator::kAllocAlignment; 88 } 89 90 protected: 91 std::unique_ptr<char[]> mem_segment_; 92 std::unique_ptr<PersistentMemoryAllocator> allocator_; 93 }; 94 95 TEST_F(PersistentMemoryAllocatorTest, AllocateAndIterate) { 96 allocator_->CreateTrackingHistograms(allocator_->Name()); 97 98 std::string base_name(TEST_NAME); 99 EXPECT_EQ(TEST_ID, allocator_->Id()); 100 EXPECT_TRUE(allocator_->used_histogram_); 101 EXPECT_EQ("UMA.PersistentAllocator." + base_name + ".UsedPct", 102 allocator_->used_histogram_->histogram_name()); 103 EXPECT_EQ(PersistentMemoryAllocator::MEMORY_INITIALIZED, 104 allocator_->GetMemoryState()); 105 106 // Get base memory info for later comparison. 107 PersistentMemoryAllocator::MemoryInfo meminfo0; 108 allocator_->GetMemoryInfo(&meminfo0); 109 EXPECT_EQ(TEST_MEMORY_SIZE, meminfo0.total); 110 EXPECT_GT(meminfo0.total, meminfo0.free); 111 112 // Validate allocation of test object and make sure it can be referenced 113 // and all metadata looks correct. 114 TestObject1* obj1 = allocator_->New<TestObject1>(); 115 ASSERT_TRUE(obj1); 116 Reference block1 = allocator_->GetAsReference(obj1); 117 ASSERT_NE(0U, block1); 118 EXPECT_NE(nullptr, allocator_->GetAsObject<TestObject1>(block1)); 119 EXPECT_EQ(nullptr, allocator_->GetAsObject<TestObject2>(block1)); 120 EXPECT_LE(sizeof(TestObject1), allocator_->GetAllocSize(block1)); 121 EXPECT_GT(sizeof(TestObject1) + kAllocAlignment, 122 allocator_->GetAllocSize(block1)); 123 PersistentMemoryAllocator::MemoryInfo meminfo1; 124 allocator_->GetMemoryInfo(&meminfo1); 125 EXPECT_EQ(meminfo0.total, meminfo1.total); 126 EXPECT_GT(meminfo0.free, meminfo1.free); 127 128 // Verify that pointers can be turned back into references and that invalid 129 // addresses return null. 130 char* memory1 = allocator_->GetAsArray<char>(block1, 1, 1); 131 ASSERT_TRUE(memory1); 132 EXPECT_EQ(block1, allocator_->GetAsReference(memory1, 0)); 133 EXPECT_EQ(block1, allocator_->GetAsReference(memory1, 1)); 134 EXPECT_EQ(0U, allocator_->GetAsReference(memory1, 2)); 135 EXPECT_EQ(0U, allocator_->GetAsReference(memory1 + 1, 0)); 136 EXPECT_EQ(0U, allocator_->GetAsReference(memory1 + 16, 0)); 137 EXPECT_EQ(0U, allocator_->GetAsReference(nullptr, 0)); 138 EXPECT_EQ(0U, allocator_->GetAsReference(&base_name, 0)); 139 140 // Ensure that the test-object can be made iterable. 141 PersistentMemoryAllocator::Iterator iter1a(allocator_.get()); 142 EXPECT_EQ(0U, iter1a.GetLast()); 143 uint32_t type; 144 EXPECT_EQ(0U, iter1a.GetNext(&type)); 145 allocator_->MakeIterable(block1); 146 EXPECT_EQ(block1, iter1a.GetNext(&type)); 147 EXPECT_EQ(1U, type); 148 EXPECT_EQ(block1, iter1a.GetLast()); 149 EXPECT_EQ(0U, iter1a.GetNext(&type)); 150 EXPECT_EQ(block1, iter1a.GetLast()); 151 152 // Create second test-object and ensure everything is good and it cannot 153 // be confused with test-object of another type. 154 TestObject2* obj2 = allocator_->New<TestObject2>(); 155 ASSERT_TRUE(obj2); 156 Reference block2 = allocator_->GetAsReference(obj2); 157 ASSERT_NE(0U, block2); 158 EXPECT_NE(nullptr, allocator_->GetAsObject<TestObject2>(block2)); 159 EXPECT_EQ(nullptr, allocator_->GetAsObject<TestObject1>(block2)); 160 EXPECT_LE(sizeof(TestObject2), allocator_->GetAllocSize(block2)); 161 EXPECT_GT(sizeof(TestObject2) + kAllocAlignment, 162 allocator_->GetAllocSize(block2)); 163 PersistentMemoryAllocator::MemoryInfo meminfo2; 164 allocator_->GetMemoryInfo(&meminfo2); 165 EXPECT_EQ(meminfo1.total, meminfo2.total); 166 EXPECT_GT(meminfo1.free, meminfo2.free); 167 168 // Ensure that second test-object can also be made iterable. 169 allocator_->MakeIterable(obj2); 170 EXPECT_EQ(block2, iter1a.GetNext(&type)); 171 EXPECT_EQ(2U, type); 172 EXPECT_EQ(block2, iter1a.GetLast()); 173 EXPECT_EQ(0U, iter1a.GetNext(&type)); 174 EXPECT_EQ(block2, iter1a.GetLast()); 175 176 // Check that the iterator can be reset to the beginning. 177 iter1a.Reset(); 178 EXPECT_EQ(0U, iter1a.GetLast()); 179 EXPECT_EQ(block1, iter1a.GetNext(&type)); 180 EXPECT_EQ(block1, iter1a.GetLast()); 181 EXPECT_EQ(block2, iter1a.GetNext(&type)); 182 EXPECT_EQ(block2, iter1a.GetLast()); 183 EXPECT_EQ(0U, iter1a.GetNext(&type)); 184 185 // Check that the iterator can be reset to an arbitrary location. 186 iter1a.Reset(block1); 187 EXPECT_EQ(block1, iter1a.GetLast()); 188 EXPECT_EQ(block2, iter1a.GetNext(&type)); 189 EXPECT_EQ(block2, iter1a.GetLast()); 190 EXPECT_EQ(0U, iter1a.GetNext(&type)); 191 192 // Check that iteration can begin after an arbitrary location. 193 PersistentMemoryAllocator::Iterator iter1b(allocator_.get(), block1); 194 EXPECT_EQ(block2, iter1b.GetNext(&type)); 195 EXPECT_EQ(0U, iter1b.GetNext(&type)); 196 197 // Ensure nothing has gone noticably wrong. 198 EXPECT_FALSE(allocator_->IsFull()); 199 EXPECT_FALSE(allocator_->IsCorrupt()); 200 201 // Check the internal histogram record of used memory. 202 allocator_->UpdateTrackingHistograms(); 203 std::unique_ptr<HistogramSamples> used_samples( 204 allocator_->used_histogram_->SnapshotSamples()); 205 EXPECT_TRUE(used_samples); 206 EXPECT_EQ(1, used_samples->TotalCount()); 207 208 // Check that an object's type can be changed. 209 EXPECT_EQ(2U, allocator_->GetType(block2)); 210 allocator_->ChangeType(block2, 3, 2, false); 211 EXPECT_EQ(3U, allocator_->GetType(block2)); 212 allocator_->New<TestObject2>(block2, 3, false); 213 EXPECT_EQ(2U, allocator_->GetType(block2)); 214 215 // Create second allocator (read/write) using the same memory segment. 216 std::unique_ptr<PersistentMemoryAllocator> allocator2( 217 new PersistentMemoryAllocator(mem_segment_.get(), TEST_MEMORY_SIZE, 218 TEST_MEMORY_PAGE, 0, "", false)); 219 EXPECT_EQ(TEST_ID, allocator2->Id()); 220 EXPECT_FALSE(allocator2->used_histogram_); 221 222 // Ensure that iteration and access through second allocator works. 223 PersistentMemoryAllocator::Iterator iter2(allocator2.get()); 224 EXPECT_EQ(block1, iter2.GetNext(&type)); 225 EXPECT_EQ(block2, iter2.GetNext(&type)); 226 EXPECT_EQ(0U, iter2.GetNext(&type)); 227 EXPECT_NE(nullptr, allocator2->GetAsObject<TestObject1>(block1)); 228 EXPECT_NE(nullptr, allocator2->GetAsObject<TestObject2>(block2)); 229 230 // Create a third allocator (read-only) using the same memory segment. 231 std::unique_ptr<const PersistentMemoryAllocator> allocator3( 232 new PersistentMemoryAllocator(mem_segment_.get(), TEST_MEMORY_SIZE, 233 TEST_MEMORY_PAGE, 0, "", true)); 234 EXPECT_EQ(TEST_ID, allocator3->Id()); 235 EXPECT_FALSE(allocator3->used_histogram_); 236 237 // Ensure that iteration and access through third allocator works. 238 PersistentMemoryAllocator::Iterator iter3(allocator3.get()); 239 EXPECT_EQ(block1, iter3.GetNext(&type)); 240 EXPECT_EQ(block2, iter3.GetNext(&type)); 241 EXPECT_EQ(0U, iter3.GetNext(&type)); 242 EXPECT_NE(nullptr, allocator3->GetAsObject<TestObject1>(block1)); 243 EXPECT_NE(nullptr, allocator3->GetAsObject<TestObject2>(block2)); 244 245 // Ensure that GetNextOfType works. 246 PersistentMemoryAllocator::Iterator iter1c(allocator_.get()); 247 EXPECT_EQ(block2, iter1c.GetNextOfType<TestObject2>()); 248 EXPECT_EQ(0U, iter1c.GetNextOfType(2)); 249 250 // Ensure that GetNextOfObject works. 251 PersistentMemoryAllocator::Iterator iter1d(allocator_.get()); 252 EXPECT_EQ(obj2, iter1d.GetNextOfObject<TestObject2>()); 253 EXPECT_EQ(nullptr, iter1d.GetNextOfObject<TestObject2>()); 254 255 // Ensure that deleting an object works. 256 allocator_->Delete(obj2); 257 PersistentMemoryAllocator::Iterator iter1z(allocator_.get()); 258 EXPECT_EQ(nullptr, iter1z.GetNextOfObject<TestObject2>()); 259 260 // Ensure that the memory state can be set. 261 allocator_->SetMemoryState(PersistentMemoryAllocator::MEMORY_DELETED); 262 EXPECT_EQ(PersistentMemoryAllocator::MEMORY_DELETED, 263 allocator_->GetMemoryState()); 264 } 265 266 TEST_F(PersistentMemoryAllocatorTest, PageTest) { 267 // This allocation will go into the first memory page. 268 Reference block1 = allocator_->Allocate(TEST_MEMORY_PAGE / 2, 1); 269 EXPECT_LT(0U, block1); 270 EXPECT_GT(TEST_MEMORY_PAGE, block1); 271 272 // This allocation won't fit in same page as previous block. 273 Reference block2 = 274 allocator_->Allocate(TEST_MEMORY_PAGE - 2 * kAllocAlignment, 2); 275 EXPECT_EQ(TEST_MEMORY_PAGE, block2); 276 277 // This allocation will also require a new page. 278 Reference block3 = allocator_->Allocate(2 * kAllocAlignment + 99, 3); 279 EXPECT_EQ(2U * TEST_MEMORY_PAGE, block3); 280 } 281 282 // A simple thread that takes an allocator and repeatedly allocates random- 283 // sized chunks from it until no more can be done. 284 class AllocatorThread : public SimpleThread { 285 public: 286 AllocatorThread(const std::string& name, 287 void* base, 288 uint32_t size, 289 uint32_t page_size) 290 : SimpleThread(name, Options()), 291 count_(0), 292 iterable_(0), 293 allocator_(base, size, page_size, 0, std::string(), false) {} 294 295 void Run() override { 296 for (;;) { 297 uint32_t size = RandInt(1, 99); 298 uint32_t type = RandInt(100, 999); 299 Reference block = allocator_.Allocate(size, type); 300 if (!block) 301 break; 302 303 count_++; 304 if (RandInt(0, 1)) { 305 allocator_.MakeIterable(block); 306 iterable_++; 307 } 308 } 309 } 310 311 unsigned iterable() { return iterable_; } 312 unsigned count() { return count_; } 313 314 private: 315 unsigned count_; 316 unsigned iterable_; 317 PersistentMemoryAllocator allocator_; 318 }; 319 320 // Test parallel allocation/iteration and ensure consistency across all 321 // instances. 322 TEST_F(PersistentMemoryAllocatorTest, ParallelismTest) { 323 void* memory = mem_segment_.get(); 324 AllocatorThread t1("t1", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 325 AllocatorThread t2("t2", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 326 AllocatorThread t3("t3", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 327 AllocatorThread t4("t4", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 328 AllocatorThread t5("t5", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 329 330 t1.Start(); 331 t2.Start(); 332 t3.Start(); 333 t4.Start(); 334 t5.Start(); 335 336 unsigned last_count = 0; 337 do { 338 unsigned count = CountIterables(); 339 EXPECT_LE(last_count, count); 340 } while (!allocator_->IsCorrupt() && !allocator_->IsFull()); 341 342 t1.Join(); 343 t2.Join(); 344 t3.Join(); 345 t4.Join(); 346 t5.Join(); 347 348 EXPECT_FALSE(allocator_->IsCorrupt()); 349 EXPECT_TRUE(allocator_->IsFull()); 350 EXPECT_EQ(CountIterables(), 351 t1.iterable() + t2.iterable() + t3.iterable() + t4.iterable() + 352 t5.iterable()); 353 } 354 355 // A simple thread that counts objects by iterating through an allocator. 356 class CounterThread : public SimpleThread { 357 public: 358 CounterThread(const std::string& name, 359 PersistentMemoryAllocator::Iterator* iterator, 360 Lock* lock, 361 ConditionVariable* condition, 362 bool* wake_up) 363 : SimpleThread(name, Options()), 364 iterator_(iterator), 365 lock_(lock), 366 condition_(condition), 367 count_(0), 368 wake_up_(wake_up) {} 369 370 void Run() override { 371 // Wait so all threads can start at approximately the same time. 372 // Best performance comes from releasing a single worker which then 373 // releases the next, etc., etc. 374 { 375 AutoLock autolock(*lock_); 376 377 // Before calling Wait(), make sure that the wake up condition 378 // has not already passed. Also, since spurious signal events 379 // are possible, check the condition in a while loop to make 380 // sure that the wake up condition is met when this thread 381 // returns from the Wait(). 382 // See usage comments in src/base/synchronization/condition_variable.h. 383 while (!*wake_up_) { 384 condition_->Wait(); 385 condition_->Signal(); 386 } 387 } 388 389 uint32_t type; 390 while (iterator_->GetNext(&type) != 0) { 391 ++count_; 392 } 393 } 394 395 unsigned count() { return count_; } 396 397 private: 398 PersistentMemoryAllocator::Iterator* iterator_; 399 Lock* lock_; 400 ConditionVariable* condition_; 401 unsigned count_; 402 bool* wake_up_; 403 404 DISALLOW_COPY_AND_ASSIGN(CounterThread); 405 }; 406 407 // Ensure that parallel iteration returns the same number of objects as 408 // single-threaded iteration. 409 TEST_F(PersistentMemoryAllocatorTest, IteratorParallelismTest) { 410 // Fill the memory segment with random allocations. 411 unsigned iterable_count = 0; 412 for (;;) { 413 uint32_t size = RandInt(1, 99); 414 uint32_t type = RandInt(100, 999); 415 Reference block = allocator_->Allocate(size, type); 416 if (!block) 417 break; 418 allocator_->MakeIterable(block); 419 ++iterable_count; 420 } 421 EXPECT_FALSE(allocator_->IsCorrupt()); 422 EXPECT_TRUE(allocator_->IsFull()); 423 EXPECT_EQ(iterable_count, CountIterables()); 424 425 PersistentMemoryAllocator::Iterator iter(allocator_.get()); 426 Lock lock; 427 ConditionVariable condition(&lock); 428 bool wake_up = false; 429 430 CounterThread t1("t1", &iter, &lock, &condition, &wake_up); 431 CounterThread t2("t2", &iter, &lock, &condition, &wake_up); 432 CounterThread t3("t3", &iter, &lock, &condition, &wake_up); 433 CounterThread t4("t4", &iter, &lock, &condition, &wake_up); 434 CounterThread t5("t5", &iter, &lock, &condition, &wake_up); 435 436 t1.Start(); 437 t2.Start(); 438 t3.Start(); 439 t4.Start(); 440 t5.Start(); 441 442 // Take the lock and set the wake up condition to true. This helps to 443 // avoid a race condition where the Signal() event is called before 444 // all the threads have reached the Wait() and thus never get woken up. 445 { 446 AutoLock autolock(lock); 447 wake_up = true; 448 } 449 450 // This will release all the waiting threads. 451 condition.Signal(); 452 453 t1.Join(); 454 t2.Join(); 455 t3.Join(); 456 t4.Join(); 457 t5.Join(); 458 459 EXPECT_EQ(iterable_count, 460 t1.count() + t2.count() + t3.count() + t4.count() + t5.count()); 461 462 #if 0 463 // These ensure that the threads don't run sequentially. It shouldn't be 464 // enabled in general because it could lead to a flaky test if it happens 465 // simply by chance but it is useful during development to ensure that the 466 // test is working correctly. 467 EXPECT_NE(iterable_count, t1.count()); 468 EXPECT_NE(iterable_count, t2.count()); 469 EXPECT_NE(iterable_count, t3.count()); 470 EXPECT_NE(iterable_count, t4.count()); 471 EXPECT_NE(iterable_count, t5.count()); 472 #endif 473 } 474 475 // This test doesn't verify anything other than it doesn't crash. Its goal 476 // is to find coding errors that aren't otherwise tested for, much like a 477 // "fuzzer" would. 478 // This test is suppsoed to fail on TSAN bot (crbug.com/579867). 479 #if defined(THREAD_SANITIZER) 480 #define MAYBE_CorruptionTest DISABLED_CorruptionTest 481 #else 482 #define MAYBE_CorruptionTest CorruptionTest 483 #endif 484 TEST_F(PersistentMemoryAllocatorTest, MAYBE_CorruptionTest) { 485 char* memory = mem_segment_.get(); 486 AllocatorThread t1("t1", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 487 AllocatorThread t2("t2", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 488 AllocatorThread t3("t3", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 489 AllocatorThread t4("t4", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 490 AllocatorThread t5("t5", memory, TEST_MEMORY_SIZE, TEST_MEMORY_PAGE); 491 492 t1.Start(); 493 t2.Start(); 494 t3.Start(); 495 t4.Start(); 496 t5.Start(); 497 498 do { 499 size_t offset = RandInt(0, TEST_MEMORY_SIZE - 1); 500 char value = RandInt(0, 255); 501 memory[offset] = value; 502 } while (!allocator_->IsCorrupt() && !allocator_->IsFull()); 503 504 t1.Join(); 505 t2.Join(); 506 t3.Join(); 507 t4.Join(); 508 t5.Join(); 509 510 CountIterables(); 511 } 512 513 // Attempt to cause crashes or loops by expressly creating dangerous conditions. 514 TEST_F(PersistentMemoryAllocatorTest, MaliciousTest) { 515 Reference block1 = allocator_->Allocate(sizeof(TestObject1), 1); 516 Reference block2 = allocator_->Allocate(sizeof(TestObject1), 2); 517 Reference block3 = allocator_->Allocate(sizeof(TestObject1), 3); 518 Reference block4 = allocator_->Allocate(sizeof(TestObject1), 3); 519 Reference block5 = allocator_->Allocate(sizeof(TestObject1), 3); 520 allocator_->MakeIterable(block1); 521 allocator_->MakeIterable(block2); 522 allocator_->MakeIterable(block3); 523 allocator_->MakeIterable(block4); 524 allocator_->MakeIterable(block5); 525 EXPECT_EQ(5U, CountIterables()); 526 EXPECT_FALSE(allocator_->IsCorrupt()); 527 528 // Create loop in iterable list and ensure it doesn't hang. The return value 529 // from CountIterables() in these cases is unpredictable. If there is a 530 // failure, the call will hang and the test killed for taking too long. 531 uint32_t* header4 = (uint32_t*)(mem_segment_.get() + block4); 532 EXPECT_EQ(block5, header4[3]); 533 header4[3] = block4; 534 CountIterables(); // loop: 1-2-3-4-4 535 EXPECT_TRUE(allocator_->IsCorrupt()); 536 537 // Test where loop goes back to previous block. 538 header4[3] = block3; 539 CountIterables(); // loop: 1-2-3-4-3 540 541 // Test where loop goes back to the beginning. 542 header4[3] = block1; 543 CountIterables(); // loop: 1-2-3-4-1 544 } 545 546 547 //----- LocalPersistentMemoryAllocator ----------------------------------------- 548 549 TEST(LocalPersistentMemoryAllocatorTest, CreationTest) { 550 LocalPersistentMemoryAllocator allocator(TEST_MEMORY_SIZE, 42, ""); 551 EXPECT_EQ(42U, allocator.Id()); 552 EXPECT_NE(0U, allocator.Allocate(24, 1)); 553 EXPECT_FALSE(allocator.IsFull()); 554 EXPECT_FALSE(allocator.IsCorrupt()); 555 } 556 557 558 //----- SharedPersistentMemoryAllocator ---------------------------------------- 559 560 TEST(SharedPersistentMemoryAllocatorTest, CreationTest) { 561 SharedMemoryHandle shared_handle_1; 562 SharedMemoryHandle shared_handle_2; 563 564 PersistentMemoryAllocator::MemoryInfo meminfo1; 565 Reference r123, r456, r789; 566 { 567 std::unique_ptr<SharedMemory> shmem1(new SharedMemory()); 568 ASSERT_TRUE(shmem1->CreateAndMapAnonymous(TEST_MEMORY_SIZE)); 569 SharedPersistentMemoryAllocator local(std::move(shmem1), TEST_ID, "", 570 false); 571 EXPECT_FALSE(local.IsReadonly()); 572 r123 = local.Allocate(123, 123); 573 r456 = local.Allocate(456, 456); 574 r789 = local.Allocate(789, 789); 575 local.MakeIterable(r123); 576 local.ChangeType(r456, 654, 456, false); 577 local.MakeIterable(r789); 578 local.GetMemoryInfo(&meminfo1); 579 EXPECT_FALSE(local.IsFull()); 580 EXPECT_FALSE(local.IsCorrupt()); 581 582 ASSERT_TRUE(local.shared_memory()->ShareToProcess(GetCurrentProcessHandle(), 583 &shared_handle_1)); 584 ASSERT_TRUE(local.shared_memory()->ShareToProcess(GetCurrentProcessHandle(), 585 &shared_handle_2)); 586 } 587 588 // Read-only test. 589 std::unique_ptr<SharedMemory> shmem2(new SharedMemory(shared_handle_1, 590 /*readonly=*/true)); 591 ASSERT_TRUE(shmem2->Map(TEST_MEMORY_SIZE)); 592 593 SharedPersistentMemoryAllocator shalloc2(std::move(shmem2), 0, "", true); 594 EXPECT_TRUE(shalloc2.IsReadonly()); 595 EXPECT_EQ(TEST_ID, shalloc2.Id()); 596 EXPECT_FALSE(shalloc2.IsFull()); 597 EXPECT_FALSE(shalloc2.IsCorrupt()); 598 599 PersistentMemoryAllocator::Iterator iter2(&shalloc2); 600 uint32_t type; 601 EXPECT_EQ(r123, iter2.GetNext(&type)); 602 EXPECT_EQ(r789, iter2.GetNext(&type)); 603 EXPECT_EQ(0U, iter2.GetNext(&type)); 604 605 EXPECT_EQ(123U, shalloc2.GetType(r123)); 606 EXPECT_EQ(654U, shalloc2.GetType(r456)); 607 EXPECT_EQ(789U, shalloc2.GetType(r789)); 608 609 PersistentMemoryAllocator::MemoryInfo meminfo2; 610 shalloc2.GetMemoryInfo(&meminfo2); 611 EXPECT_EQ(meminfo1.total, meminfo2.total); 612 EXPECT_EQ(meminfo1.free, meminfo2.free); 613 614 // Read/write test. 615 std::unique_ptr<SharedMemory> shmem3(new SharedMemory(shared_handle_2, 616 /*readonly=*/false)); 617 ASSERT_TRUE(shmem3->Map(TEST_MEMORY_SIZE)); 618 619 SharedPersistentMemoryAllocator shalloc3(std::move(shmem3), 0, "", false); 620 EXPECT_FALSE(shalloc3.IsReadonly()); 621 EXPECT_EQ(TEST_ID, shalloc3.Id()); 622 EXPECT_FALSE(shalloc3.IsFull()); 623 EXPECT_FALSE(shalloc3.IsCorrupt()); 624 625 PersistentMemoryAllocator::Iterator iter3(&shalloc3); 626 EXPECT_EQ(r123, iter3.GetNext(&type)); 627 EXPECT_EQ(r789, iter3.GetNext(&type)); 628 EXPECT_EQ(0U, iter3.GetNext(&type)); 629 630 EXPECT_EQ(123U, shalloc3.GetType(r123)); 631 EXPECT_EQ(654U, shalloc3.GetType(r456)); 632 EXPECT_EQ(789U, shalloc3.GetType(r789)); 633 634 PersistentMemoryAllocator::MemoryInfo meminfo3; 635 shalloc3.GetMemoryInfo(&meminfo3); 636 EXPECT_EQ(meminfo1.total, meminfo3.total); 637 EXPECT_EQ(meminfo1.free, meminfo3.free); 638 639 // Interconnectivity test. 640 Reference obj = shalloc3.Allocate(42, 42); 641 ASSERT_TRUE(obj); 642 shalloc3.MakeIterable(obj); 643 EXPECT_EQ(obj, iter2.GetNext(&type)); 644 EXPECT_EQ(42U, type); 645 646 // Clear-on-change test. 647 Reference data_ref = shalloc3.Allocate(sizeof(int) * 4, 911); 648 int* data = shalloc3.GetAsArray<int>(data_ref, 911, 4); 649 ASSERT_TRUE(data); 650 data[0] = 0; 651 data[1] = 1; 652 data[2] = 2; 653 data[3] = 3; 654 ASSERT_TRUE(shalloc3.ChangeType(data_ref, 119, 911, false)); 655 EXPECT_EQ(0, data[0]); 656 EXPECT_EQ(1, data[1]); 657 EXPECT_EQ(2, data[2]); 658 EXPECT_EQ(3, data[3]); 659 ASSERT_TRUE(shalloc3.ChangeType(data_ref, 191, 119, true)); 660 EXPECT_EQ(0, data[0]); 661 EXPECT_EQ(0, data[1]); 662 EXPECT_EQ(0, data[2]); 663 EXPECT_EQ(0, data[3]); 664 } 665 666 667 #if !defined(OS_NACL) 668 //----- FilePersistentMemoryAllocator ------------------------------------------ 669 670 TEST(FilePersistentMemoryAllocatorTest, CreationTest) { 671 ScopedTempDir temp_dir; 672 ASSERT_TRUE(temp_dir.CreateUniqueTempDir()); 673 FilePath file_path = temp_dir.GetPath().AppendASCII("persistent_memory"); 674 675 PersistentMemoryAllocator::MemoryInfo meminfo1; 676 Reference r123, r456, r789; 677 { 678 LocalPersistentMemoryAllocator local(TEST_MEMORY_SIZE, TEST_ID, ""); 679 EXPECT_FALSE(local.IsReadonly()); 680 r123 = local.Allocate(123, 123); 681 r456 = local.Allocate(456, 456); 682 r789 = local.Allocate(789, 789); 683 local.MakeIterable(r123); 684 local.ChangeType(r456, 654, 456, false); 685 local.MakeIterable(r789); 686 local.GetMemoryInfo(&meminfo1); 687 EXPECT_FALSE(local.IsFull()); 688 EXPECT_FALSE(local.IsCorrupt()); 689 690 File writer(file_path, File::FLAG_CREATE | File::FLAG_WRITE); 691 ASSERT_TRUE(writer.IsValid()); 692 writer.Write(0, (const char*)local.data(), local.used()); 693 } 694 695 std::unique_ptr<MemoryMappedFile> mmfile(new MemoryMappedFile()); 696 mmfile->Initialize(file_path); 697 EXPECT_TRUE(mmfile->IsValid()); 698 const size_t mmlength = mmfile->length(); 699 EXPECT_GE(meminfo1.total, mmlength); 700 701 FilePersistentMemoryAllocator file(std::move(mmfile), 0, 0, "", false); 702 EXPECT_FALSE(file.IsReadonly()); 703 EXPECT_EQ(TEST_ID, file.Id()); 704 EXPECT_FALSE(file.IsFull()); 705 EXPECT_FALSE(file.IsCorrupt()); 706 707 PersistentMemoryAllocator::Iterator iter(&file); 708 uint32_t type; 709 EXPECT_EQ(r123, iter.GetNext(&type)); 710 EXPECT_EQ(r789, iter.GetNext(&type)); 711 EXPECT_EQ(0U, iter.GetNext(&type)); 712 713 EXPECT_EQ(123U, file.GetType(r123)); 714 EXPECT_EQ(654U, file.GetType(r456)); 715 EXPECT_EQ(789U, file.GetType(r789)); 716 717 PersistentMemoryAllocator::MemoryInfo meminfo2; 718 file.GetMemoryInfo(&meminfo2); 719 EXPECT_GE(meminfo1.total, meminfo2.total); 720 EXPECT_GE(meminfo1.free, meminfo2.free); 721 EXPECT_EQ(mmlength, meminfo2.total); 722 EXPECT_EQ(0U, meminfo2.free); 723 724 // There's no way of knowing if Flush actually does anything but at least 725 // verify that it runs without CHECK violations. 726 file.Flush(false); 727 file.Flush(true); 728 } 729 730 TEST(FilePersistentMemoryAllocatorTest, ExtendTest) { 731 ScopedTempDir temp_dir; 732 ASSERT_TRUE(temp_dir.CreateUniqueTempDir()); 733 FilePath file_path = temp_dir.GetPath().AppendASCII("extend_test"); 734 MemoryMappedFile::Region region = {0, 16 << 10}; // 16KiB maximum size. 735 736 // Start with a small but valid file of persistent data. 737 ASSERT_FALSE(PathExists(file_path)); 738 { 739 LocalPersistentMemoryAllocator local(TEST_MEMORY_SIZE, TEST_ID, ""); 740 local.Allocate(1, 1); 741 local.Allocate(11, 11); 742 743 File writer(file_path, File::FLAG_CREATE | File::FLAG_WRITE); 744 ASSERT_TRUE(writer.IsValid()); 745 writer.Write(0, (const char*)local.data(), local.used()); 746 } 747 ASSERT_TRUE(PathExists(file_path)); 748 int64_t before_size; 749 ASSERT_TRUE(GetFileSize(file_path, &before_size)); 750 751 // Map it as an extendable read/write file and append to it. 752 { 753 std::unique_ptr<MemoryMappedFile> mmfile(new MemoryMappedFile()); 754 mmfile->Initialize( 755 File(file_path, File::FLAG_OPEN | File::FLAG_READ | File::FLAG_WRITE), 756 region, MemoryMappedFile::READ_WRITE_EXTEND); 757 FilePersistentMemoryAllocator allocator(std::move(mmfile), region.size, 0, 758 "", false); 759 EXPECT_EQ(static_cast<size_t>(before_size), allocator.used()); 760 761 allocator.Allocate(111, 111); 762 EXPECT_LT(static_cast<size_t>(before_size), allocator.used()); 763 } 764 765 // Validate that append worked. 766 int64_t after_size; 767 ASSERT_TRUE(GetFileSize(file_path, &after_size)); 768 EXPECT_LT(before_size, after_size); 769 770 // Verify that it's still an acceptable file. 771 { 772 std::unique_ptr<MemoryMappedFile> mmfile(new MemoryMappedFile()); 773 mmfile->Initialize( 774 File(file_path, File::FLAG_OPEN | File::FLAG_READ | File::FLAG_WRITE), 775 region, MemoryMappedFile::READ_WRITE_EXTEND); 776 EXPECT_TRUE(FilePersistentMemoryAllocator::IsFileAcceptable(*mmfile, true)); 777 EXPECT_TRUE( 778 FilePersistentMemoryAllocator::IsFileAcceptable(*mmfile, false)); 779 } 780 } 781 782 TEST(FilePersistentMemoryAllocatorTest, AcceptableTest) { 783 const uint32_t kAllocAlignment = 784 PersistentMemoryAllocatorTest::GetAllocAlignment(); 785 ScopedTempDir temp_dir; 786 ASSERT_TRUE(temp_dir.CreateUniqueTempDir()); 787 788 LocalPersistentMemoryAllocator local(TEST_MEMORY_SIZE, TEST_ID, ""); 789 local.MakeIterable(local.Allocate(1, 1)); 790 local.MakeIterable(local.Allocate(11, 11)); 791 const size_t minsize = local.used(); 792 std::unique_ptr<char[]> garbage(new char[minsize]); 793 RandBytes(garbage.get(), minsize); 794 795 std::unique_ptr<MemoryMappedFile> mmfile; 796 char filename[100]; 797 for (size_t filesize = minsize; filesize > 0; --filesize) { 798 strings::SafeSPrintf(filename, "memory_%d_A", filesize); 799 FilePath file_path = temp_dir.GetPath().AppendASCII(filename); 800 ASSERT_FALSE(PathExists(file_path)); 801 { 802 File writer(file_path, File::FLAG_CREATE | File::FLAG_WRITE); 803 ASSERT_TRUE(writer.IsValid()); 804 writer.Write(0, (const char*)local.data(), filesize); 805 } 806 ASSERT_TRUE(PathExists(file_path)); 807 808 // Request read/write access for some sizes that are a multple of the 809 // allocator's alignment size. The allocator is strict about file size 810 // being a multiple of its internal alignment when doing read/write access. 811 const bool read_only = (filesize % (2 * kAllocAlignment)) != 0; 812 const uint32_t file_flags = 813 File::FLAG_OPEN | File::FLAG_READ | (read_only ? 0 : File::FLAG_WRITE); 814 const MemoryMappedFile::Access map_access = 815 read_only ? MemoryMappedFile::READ_ONLY : MemoryMappedFile::READ_WRITE; 816 817 mmfile.reset(new MemoryMappedFile()); 818 mmfile->Initialize(File(file_path, file_flags), map_access); 819 EXPECT_EQ(filesize, mmfile->length()); 820 if (FilePersistentMemoryAllocator::IsFileAcceptable(*mmfile, read_only)) { 821 // Make sure construction doesn't crash. It will, however, cause 822 // error messages warning about about a corrupted memory segment. 823 FilePersistentMemoryAllocator allocator(std::move(mmfile), 0, 0, "", 824 read_only); 825 // Also make sure that iteration doesn't crash. 826 PersistentMemoryAllocator::Iterator iter(&allocator); 827 uint32_t type_id; 828 Reference ref; 829 while ((ref = iter.GetNext(&type_id)) != 0) { 830 const char* data = allocator.GetAsArray<char>( 831 ref, 0, PersistentMemoryAllocator::kSizeAny); 832 uint32_t type = allocator.GetType(ref); 833 size_t size = allocator.GetAllocSize(ref); 834 // Ensure compiler can't optimize-out above variables. 835 (void)data; 836 (void)type; 837 (void)size; 838 } 839 840 // Ensure that short files are detected as corrupt and full files are not. 841 EXPECT_EQ(filesize != minsize, allocator.IsCorrupt()); 842 } else { 843 // For filesize >= minsize, the file must be acceptable. This 844 // else clause (file-not-acceptable) should be reached only if 845 // filesize < minsize. 846 EXPECT_LT(filesize, minsize); 847 } 848 849 strings::SafeSPrintf(filename, "memory_%d_B", filesize); 850 file_path = temp_dir.GetPath().AppendASCII(filename); 851 ASSERT_FALSE(PathExists(file_path)); 852 { 853 File writer(file_path, File::FLAG_CREATE | File::FLAG_WRITE); 854 ASSERT_TRUE(writer.IsValid()); 855 writer.Write(0, (const char*)garbage.get(), filesize); 856 } 857 ASSERT_TRUE(PathExists(file_path)); 858 859 mmfile.reset(new MemoryMappedFile()); 860 mmfile->Initialize(File(file_path, file_flags), map_access); 861 EXPECT_EQ(filesize, mmfile->length()); 862 if (FilePersistentMemoryAllocator::IsFileAcceptable(*mmfile, read_only)) { 863 // Make sure construction doesn't crash. It will, however, cause 864 // error messages warning about about a corrupted memory segment. 865 FilePersistentMemoryAllocator allocator(std::move(mmfile), 0, 0, "", 866 read_only); 867 EXPECT_TRUE(allocator.IsCorrupt()); // Garbage data so it should be. 868 } else { 869 // For filesize >= minsize, the file must be acceptable. This 870 // else clause (file-not-acceptable) should be reached only if 871 // filesize < minsize. 872 EXPECT_GT(minsize, filesize); 873 } 874 } 875 } 876 #endif // !defined(OS_NACL) 877 878 } // namespace base 879