1 /* 2 * Copyright (C) 2014 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #include "swap_space.h" 18 19 #include <algorithm> 20 #include <numeric> 21 #include <sys/mman.h> 22 23 #include "base/logging.h" 24 #include "base/macros.h" 25 #include "base/mutex.h" 26 #include "thread-inl.h" 27 28 namespace art { 29 30 // The chunk size by which the swap file is increased and mapped. 31 static constexpr size_t kMininumMapSize = 16 * MB; 32 33 static constexpr bool kCheckFreeMaps = false; 34 35 template <typename FreeBySizeSet> 36 static void DumpFreeMap(const FreeBySizeSet& free_by_size) { 37 size_t last_size = static_cast<size_t>(-1); 38 for (const auto& entry : free_by_size) { 39 if (last_size != entry.first) { 40 last_size = entry.first; 41 LOG(INFO) << "Size " << last_size; 42 } 43 LOG(INFO) << " 0x" << std::hex << entry.second->Start() 44 << " size=" << std::dec << entry.second->size; 45 } 46 } 47 48 void SwapSpace::RemoveChunk(FreeBySizeSet::const_iterator free_by_size_pos) { 49 auto free_by_start_pos = free_by_size_pos->second; 50 free_by_size_.erase(free_by_size_pos); 51 free_by_start_.erase(free_by_start_pos); 52 } 53 54 inline void SwapSpace::InsertChunk(const SpaceChunk& chunk) { 55 DCHECK_NE(chunk.size, 0u); 56 auto insert_result = free_by_start_.insert(chunk); 57 DCHECK(insert_result.second); 58 free_by_size_.emplace(chunk.size, insert_result.first); 59 } 60 61 SwapSpace::SwapSpace(int fd, size_t initial_size) 62 : fd_(fd), 63 size_(0), 64 lock_("SwapSpace lock", static_cast<LockLevel>(LockLevel::kDefaultMutexLevel - 1)) { 65 // Assume that the file is unlinked. 66 67 InsertChunk(NewFileChunk(initial_size)); 68 } 69 70 SwapSpace::~SwapSpace() { 71 // Unmap all mmapped chunks. Nothing should be allocated anymore at 72 // this point, so there should be only full size chunks in free_by_start_. 73 for (const SpaceChunk& chunk : free_by_start_) { 74 if (munmap(chunk.ptr, chunk.size) != 0) { 75 PLOG(ERROR) << "Failed to unmap swap space chunk at " 76 << static_cast<const void*>(chunk.ptr) << " size=" << chunk.size; 77 } 78 } 79 // All arenas are backed by the same file. Just close the descriptor. 80 close(fd_); 81 } 82 83 template <typename FreeByStartSet, typename FreeBySizeSet> 84 static size_t CollectFree(const FreeByStartSet& free_by_start, const FreeBySizeSet& free_by_size) { 85 if (free_by_start.size() != free_by_size.size()) { 86 LOG(FATAL) << "Size: " << free_by_start.size() << " vs " << free_by_size.size(); 87 } 88 89 // Calculate over free_by_size. 90 size_t sum1 = 0; 91 for (const auto& entry : free_by_size) { 92 sum1 += entry.second->size; 93 } 94 95 // Calculate over free_by_start. 96 size_t sum2 = 0; 97 for (const auto& entry : free_by_start) { 98 sum2 += entry.size; 99 } 100 101 if (sum1 != sum2) { 102 LOG(FATAL) << "Sum: " << sum1 << " vs " << sum2; 103 } 104 return sum1; 105 } 106 107 void* SwapSpace::Alloc(size_t size) { 108 MutexLock lock(Thread::Current(), lock_); 109 size = RoundUp(size, 8U); 110 111 // Check the free list for something that fits. 112 // TODO: Smarter implementation. Global biggest chunk, ... 113 SpaceChunk old_chunk; 114 auto it = free_by_start_.empty() 115 ? free_by_size_.end() 116 : free_by_size_.lower_bound(FreeBySizeEntry { size, free_by_start_.begin() }); 117 if (it != free_by_size_.end()) { 118 old_chunk = *it->second; 119 RemoveChunk(it); 120 } else { 121 // Not a big enough free chunk, need to increase file size. 122 old_chunk = NewFileChunk(size); 123 } 124 125 void* ret = old_chunk.ptr; 126 127 if (old_chunk.size != size) { 128 // Insert the remainder. 129 SpaceChunk new_chunk = { old_chunk.ptr + size, old_chunk.size - size }; 130 InsertChunk(new_chunk); 131 } 132 133 return ret; 134 } 135 136 SwapSpace::SpaceChunk SwapSpace::NewFileChunk(size_t min_size) { 137 #if !defined(__APPLE__) 138 size_t next_part = std::max(RoundUp(min_size, kPageSize), RoundUp(kMininumMapSize, kPageSize)); 139 int result = TEMP_FAILURE_RETRY(ftruncate64(fd_, size_ + next_part)); 140 if (result != 0) { 141 PLOG(FATAL) << "Unable to increase swap file."; 142 } 143 uint8_t* ptr = reinterpret_cast<uint8_t*>( 144 mmap(nullptr, next_part, PROT_READ | PROT_WRITE, MAP_SHARED, fd_, size_)); 145 if (ptr == MAP_FAILED) { 146 LOG(ERROR) << "Unable to mmap new swap file chunk."; 147 LOG(ERROR) << "Current size: " << size_ << " requested: " << next_part << "/" << min_size; 148 LOG(ERROR) << "Free list:"; 149 DumpFreeMap(free_by_size_); 150 LOG(ERROR) << "In free list: " << CollectFree(free_by_start_, free_by_size_); 151 LOG(FATAL) << "Aborting..."; 152 } 153 size_ += next_part; 154 SpaceChunk new_chunk = {ptr, next_part}; 155 return new_chunk; 156 #else 157 UNUSED(min_size, kMininumMapSize); 158 LOG(FATAL) << "No swap file support on the Mac."; 159 UNREACHABLE(); 160 #endif 161 } 162 163 // TODO: Full coalescing. 164 void SwapSpace::Free(void* ptr, size_t size) { 165 MutexLock lock(Thread::Current(), lock_); 166 size = RoundUp(size, 8U); 167 168 size_t free_before = 0; 169 if (kCheckFreeMaps) { 170 free_before = CollectFree(free_by_start_, free_by_size_); 171 } 172 173 SpaceChunk chunk = { reinterpret_cast<uint8_t*>(ptr), size }; 174 auto it = free_by_start_.lower_bound(chunk); 175 if (it != free_by_start_.begin()) { 176 auto prev = it; 177 --prev; 178 CHECK_LE(prev->End(), chunk.Start()); 179 if (prev->End() == chunk.Start()) { 180 // Merge *prev with this chunk. 181 chunk.size += prev->size; 182 chunk.ptr -= prev->size; 183 auto erase_pos = free_by_size_.find(FreeBySizeEntry { prev->size, prev }); 184 DCHECK(erase_pos != free_by_size_.end()); 185 RemoveChunk(erase_pos); 186 // "prev" is invalidated but "it" remains valid. 187 } 188 } 189 if (it != free_by_start_.end()) { 190 CHECK_LE(chunk.End(), it->Start()); 191 if (chunk.End() == it->Start()) { 192 // Merge *it with this chunk. 193 chunk.size += it->size; 194 auto erase_pos = free_by_size_.find(FreeBySizeEntry { it->size, it }); 195 DCHECK(erase_pos != free_by_size_.end()); 196 RemoveChunk(erase_pos); 197 // "it" is invalidated but we don't need it anymore. 198 } 199 } 200 InsertChunk(chunk); 201 202 if (kCheckFreeMaps) { 203 size_t free_after = CollectFree(free_by_start_, free_by_size_); 204 205 if (free_after != free_before + size) { 206 DumpFreeMap(free_by_size_); 207 CHECK_EQ(free_after, free_before + size) << "Should be " << size << " difference from " << free_before; 208 } 209 } 210 } 211 212 } // namespace art 213