1 /* 2 * Copyright (C) 2008 Apple Inc. All rights reserved. 3 * 4 * Redistribution and use in source and binary forms, with or without 5 * modification, are permitted provided that the following conditions 6 * are met: 7 * 1. Redistributions of source code must retain the above copyright 8 * notice, this list of conditions and the following disclaimer. 9 * 2. Redistributions in binary form must reproduce the above copyright 10 * notice, this list of conditions and the following disclaimer in the 11 * documentation and/or other materials provided with the distribution. 12 * 13 * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY 14 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 15 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 16 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR 17 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, 18 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 19 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR 20 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY 21 * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 22 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 23 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 24 */ 25 26 #ifndef ExecutableAllocator_h 27 #define ExecutableAllocator_h 28 #include <stddef.h> // for ptrdiff_t 29 #include <limits> 30 #include <wtf/Assertions.h> 31 #include <wtf/PageAllocation.h> 32 #include <wtf/PassRefPtr.h> 33 #include <wtf/RefCounted.h> 34 #include <wtf/UnusedParam.h> 35 #include <wtf/Vector.h> 36 37 #if OS(IOS) 38 #include <libkern/OSCacheControl.h> 39 #include <sys/mman.h> 40 #endif 41 42 #if OS(SYMBIAN) 43 #include <e32std.h> 44 #endif 45 46 #if CPU(MIPS) && OS(LINUX) 47 #include <sys/cachectl.h> 48 #endif 49 50 #if CPU(SH4) && OS(LINUX) 51 #include <asm/cachectl.h> 52 #include <asm/unistd.h> 53 #include <sys/syscall.h> 54 #include <unistd.h> 55 #endif 56 57 #if OS(WINCE) 58 // From pkfuncs.h (private header file from the Platform Builder) 59 #define CACHE_SYNC_ALL 0x07F 60 extern "C" __declspec(dllimport) void CacheRangeFlush(LPVOID pAddr, DWORD dwLength, DWORD dwFlags); 61 #endif 62 63 #if PLATFORM(BREWMP) 64 #include <AEEIMemCache1.h> 65 #include <AEEMemCache1.bid> 66 #include <wtf/brew/RefPtrBrew.h> 67 #endif 68 69 #define JIT_ALLOCATOR_PAGE_SIZE (ExecutableAllocator::pageSize) 70 #define JIT_ALLOCATOR_LARGE_ALLOC_SIZE (ExecutableAllocator::pageSize * 4) 71 72 #if ENABLE(ASSEMBLER_WX_EXCLUSIVE) 73 #define PROTECTION_FLAGS_RW (PROT_READ | PROT_WRITE) 74 #define PROTECTION_FLAGS_RX (PROT_READ | PROT_EXEC) 75 #define EXECUTABLE_POOL_WRITABLE false 76 #else 77 #define EXECUTABLE_POOL_WRITABLE true 78 #endif 79 80 namespace JSC { 81 82 inline size_t roundUpAllocationSize(size_t request, size_t granularity) 83 { 84 if ((std::numeric_limits<size_t>::max() - granularity) <= request) 85 CRASH(); // Allocation is too large 86 87 // Round up to next page boundary 88 size_t size = request + (granularity - 1); 89 size = size & ~(granularity - 1); 90 ASSERT(size >= request); 91 return size; 92 } 93 94 } 95 96 #if ENABLE(JIT) && ENABLE(ASSEMBLER) 97 98 namespace JSC { 99 100 class ExecutablePool : public RefCounted<ExecutablePool> { 101 public: 102 #if ENABLE(EXECUTABLE_ALLOCATOR_DEMAND) 103 typedef PageAllocation Allocation; 104 #else 105 class Allocation { 106 public: 107 Allocation(void* base, size_t size) 108 : m_base(base) 109 , m_size(size) 110 { 111 } 112 void* base() { return m_base; } 113 size_t size() { return m_size; } 114 bool operator!() const { return !m_base; } 115 116 private: 117 void* m_base; 118 size_t m_size; 119 }; 120 #endif 121 typedef Vector<Allocation, 2> AllocationList; 122 123 static PassRefPtr<ExecutablePool> create(size_t n) 124 { 125 return adoptRef(new ExecutablePool(n)); 126 } 127 128 void* alloc(size_t n) 129 { 130 ASSERT(m_freePtr <= m_end); 131 132 // Round 'n' up to a multiple of word size; if all allocations are of 133 // word sized quantities, then all subsequent allocations will be aligned. 134 n = roundUpAllocationSize(n, sizeof(void*)); 135 136 if (static_cast<ptrdiff_t>(n) < (m_end - m_freePtr)) { 137 void* result = m_freePtr; 138 m_freePtr += n; 139 return result; 140 } 141 142 // Insufficient space to allocate in the existing pool 143 // so we need allocate into a new pool 144 return poolAllocate(n); 145 } 146 147 void tryShrink(void* allocation, size_t oldSize, size_t newSize) 148 { 149 if (static_cast<char*>(allocation) + oldSize != m_freePtr) 150 return; 151 m_freePtr = static_cast<char*>(allocation) + roundUpAllocationSize(newSize, sizeof(void*)); 152 } 153 154 ~ExecutablePool() 155 { 156 AllocationList::iterator end = m_pools.end(); 157 for (AllocationList::iterator ptr = m_pools.begin(); ptr != end; ++ptr) 158 ExecutablePool::systemRelease(*ptr); 159 } 160 161 size_t available() const { return (m_pools.size() > 1) ? 0 : m_end - m_freePtr; } 162 163 private: 164 static Allocation systemAlloc(size_t n); 165 static void systemRelease(Allocation& alloc); 166 167 ExecutablePool(size_t n); 168 169 void* poolAllocate(size_t n); 170 171 char* m_freePtr; 172 char* m_end; 173 AllocationList m_pools; 174 }; 175 176 class ExecutableAllocator { 177 enum ProtectionSetting { Writable, Executable }; 178 179 public: 180 static size_t pageSize; 181 ExecutableAllocator() 182 { 183 if (!pageSize) 184 intializePageSize(); 185 if (isValid()) 186 m_smallAllocationPool = ExecutablePool::create(JIT_ALLOCATOR_LARGE_ALLOC_SIZE); 187 #if !ENABLE(INTERPRETER) 188 else 189 CRASH(); 190 #endif 191 } 192 193 bool isValid() const; 194 195 static bool underMemoryPressure(); 196 197 PassRefPtr<ExecutablePool> poolForSize(size_t n) 198 { 199 // Try to fit in the existing small allocator 200 ASSERT(m_smallAllocationPool); 201 if (n < m_smallAllocationPool->available()) 202 return m_smallAllocationPool; 203 204 // If the request is large, we just provide a unshared allocator 205 if (n > JIT_ALLOCATOR_LARGE_ALLOC_SIZE) 206 return ExecutablePool::create(n); 207 208 // Create a new allocator 209 RefPtr<ExecutablePool> pool = ExecutablePool::create(JIT_ALLOCATOR_LARGE_ALLOC_SIZE); 210 211 // If the new allocator will result in more free space than in 212 // the current small allocator, then we will use it instead 213 if ((pool->available() - n) > m_smallAllocationPool->available()) 214 m_smallAllocationPool = pool; 215 return pool.release(); 216 } 217 218 #if ENABLE(ASSEMBLER_WX_EXCLUSIVE) 219 static void makeWritable(void* start, size_t size) 220 { 221 reprotectRegion(start, size, Writable); 222 } 223 224 static void makeExecutable(void* start, size_t size) 225 { 226 reprotectRegion(start, size, Executable); 227 } 228 #else 229 static void makeWritable(void*, size_t) {} 230 static void makeExecutable(void*, size_t) {} 231 #endif 232 233 234 #if CPU(X86) || CPU(X86_64) 235 static void cacheFlush(void*, size_t) 236 { 237 } 238 #elif CPU(MIPS) 239 static void cacheFlush(void* code, size_t size) 240 { 241 #if GCC_VERSION_AT_LEAST(4, 3, 0) 242 #if WTF_MIPS_ISA_REV(2) && !GCC_VERSION_AT_LEAST(4, 4, 3) 243 int lineSize; 244 asm("rdhwr %0, $1" : "=r" (lineSize)); 245 // 246 // Modify "start" and "end" to avoid GCC 4.3.0-4.4.2 bug in 247 // mips_expand_synci_loop that may execute synci one more time. 248 // "start" points to the fisrt byte of the cache line. 249 // "end" points to the last byte of the line before the last cache line. 250 // Because size is always a multiple of 4, this is safe to set 251 // "end" to the last byte. 252 // 253 intptr_t start = reinterpret_cast<intptr_t>(code) & (-lineSize); 254 intptr_t end = ((reinterpret_cast<intptr_t>(code) + size - 1) & (-lineSize)) - 1; 255 __builtin___clear_cache(reinterpret_cast<char*>(start), reinterpret_cast<char*>(end)); 256 #else 257 intptr_t end = reinterpret_cast<intptr_t>(code) + size; 258 __builtin___clear_cache(reinterpret_cast<char*>(code), reinterpret_cast<char*>(end)); 259 #endif 260 #else 261 _flush_cache(reinterpret_cast<char*>(code), size, BCACHE); 262 #endif 263 } 264 #elif CPU(ARM_THUMB2) && OS(IOS) 265 static void cacheFlush(void* code, size_t size) 266 { 267 sys_cache_control(kCacheFunctionPrepareForExecution, code, size); 268 } 269 #elif CPU(ARM_THUMB2) && OS(LINUX) 270 static void cacheFlush(void* code, size_t size) 271 { 272 asm volatile ( 273 "push {r7}\n" 274 "mov r0, %0\n" 275 "mov r1, %1\n" 276 "movw r7, #0x2\n" 277 "movt r7, #0xf\n" 278 "movs r2, #0x0\n" 279 "svc 0x0\n" 280 "pop {r7}\n" 281 : 282 : "r" (code), "r" (reinterpret_cast<char*>(code) + size) 283 : "r0", "r1", "r2"); 284 } 285 #elif OS(SYMBIAN) 286 static void cacheFlush(void* code, size_t size) 287 { 288 User::IMB_Range(code, static_cast<char*>(code) + size); 289 } 290 #elif CPU(ARM_TRADITIONAL) && OS(LINUX) && COMPILER(RVCT) 291 static __asm void cacheFlush(void* code, size_t size); 292 #elif CPU(ARM_TRADITIONAL) && OS(LINUX) && COMPILER(GCC) 293 static void cacheFlush(void* code, size_t size) 294 { 295 asm volatile ( 296 "push {r7}\n" 297 "mov r0, %0\n" 298 "mov r1, %1\n" 299 "mov r7, #0xf0000\n" 300 "add r7, r7, #0x2\n" 301 "mov r2, #0x0\n" 302 "svc 0x0\n" 303 "pop {r7}\n" 304 : 305 : "r" (code), "r" (reinterpret_cast<char*>(code) + size) 306 : "r0", "r1", "r2"); 307 } 308 #elif OS(WINCE) 309 static void cacheFlush(void* code, size_t size) 310 { 311 CacheRangeFlush(code, size, CACHE_SYNC_ALL); 312 } 313 #elif PLATFORM(BREWMP) 314 static void cacheFlush(void* code, size_t size) 315 { 316 RefPtr<IMemCache1> memCache = createRefPtrInstance<IMemCache1>(AEECLSID_MemCache1); 317 IMemCache1_ClearCache(memCache.get(), reinterpret_cast<uint32>(code), size, MEMSPACE_CACHE_FLUSH, MEMSPACE_DATACACHE); 318 IMemCache1_ClearCache(memCache.get(), reinterpret_cast<uint32>(code), size, MEMSPACE_CACHE_INVALIDATE, MEMSPACE_INSTCACHE); 319 } 320 #elif CPU(SH4) && OS(LINUX) 321 static void cacheFlush(void* code, size_t size) 322 { 323 #ifdef CACHEFLUSH_D_L2 324 syscall(__NR_cacheflush, reinterpret_cast<unsigned>(code), size, CACHEFLUSH_D_WB | CACHEFLUSH_I | CACHEFLUSH_D_L2); 325 #else 326 syscall(__NR_cacheflush, reinterpret_cast<unsigned>(code), size, CACHEFLUSH_D_WB | CACHEFLUSH_I); 327 #endif 328 } 329 #else 330 #error "The cacheFlush support is missing on this platform." 331 #endif 332 static size_t committedByteCount(); 333 334 private: 335 336 #if ENABLE(ASSEMBLER_WX_EXCLUSIVE) 337 static void reprotectRegion(void*, size_t, ProtectionSetting); 338 #endif 339 340 RefPtr<ExecutablePool> m_smallAllocationPool; 341 static void intializePageSize(); 342 }; 343 344 inline ExecutablePool::ExecutablePool(size_t n) 345 { 346 size_t allocSize = roundUpAllocationSize(n, JIT_ALLOCATOR_PAGE_SIZE); 347 Allocation mem = systemAlloc(allocSize); 348 m_pools.append(mem); 349 m_freePtr = static_cast<char*>(mem.base()); 350 if (!m_freePtr) 351 CRASH(); // Failed to allocate 352 m_end = m_freePtr + allocSize; 353 } 354 355 inline void* ExecutablePool::poolAllocate(size_t n) 356 { 357 size_t allocSize = roundUpAllocationSize(n, JIT_ALLOCATOR_PAGE_SIZE); 358 359 Allocation result = systemAlloc(allocSize); 360 if (!result.base()) 361 CRASH(); // Failed to allocate 362 363 ASSERT(m_end >= m_freePtr); 364 if ((allocSize - n) > static_cast<size_t>(m_end - m_freePtr)) { 365 // Replace allocation pool 366 m_freePtr = static_cast<char*>(result.base()) + n; 367 m_end = static_cast<char*>(result.base()) + allocSize; 368 } 369 370 m_pools.append(result); 371 return result.base(); 372 } 373 374 } 375 376 #endif // ENABLE(JIT) && ENABLE(ASSEMBLER) 377 378 #endif // !defined(ExecutableAllocator) 379