1 /* 2 * Copyright (C) 2009 University of Szeged 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY UNIVERSITY OF SZEGED ``AS IS'' AND ANY 15 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 17 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL UNIVERSITY OF SZEGED OR 18 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, 19 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 20 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR 21 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY 22 * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 23 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 24 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 25 */ 26 27 #include "config.h" 28 29 #if ENABLE(ASSEMBLER) && CPU(ARM_TRADITIONAL) 30 31 #include "ARMAssembler.h" 32 33 namespace JSC { 34 35 // Patching helpers 36 37 void ARMAssembler::patchConstantPoolLoad(void* loadAddr, void* constPoolAddr) 38 { 39 ARMWord *ldr = reinterpret_cast<ARMWord*>(loadAddr); 40 ARMWord diff = reinterpret_cast<ARMWord*>(constPoolAddr) - ldr; 41 ARMWord index = (*ldr & 0xfff) >> 1; 42 43 ASSERT(diff >= 1); 44 if (diff >= 2 || index > 0) { 45 diff = (diff + index - 2) * sizeof(ARMWord); 46 ASSERT(diff <= 0xfff); 47 *ldr = (*ldr & ~0xfff) | diff; 48 } else 49 *ldr = (*ldr & ~(0xfff | ARMAssembler::DT_UP)) | sizeof(ARMWord); 50 } 51 52 // Handle immediates 53 54 ARMWord ARMAssembler::getOp2(ARMWord imm) 55 { 56 int rol; 57 58 if (imm <= 0xff) 59 return OP2_IMM | imm; 60 61 if ((imm & 0xff000000) == 0) { 62 imm <<= 8; 63 rol = 8; 64 } 65 else { 66 imm = (imm << 24) | (imm >> 8); 67 rol = 0; 68 } 69 70 if ((imm & 0xff000000) == 0) { 71 imm <<= 8; 72 rol += 4; 73 } 74 75 if ((imm & 0xf0000000) == 0) { 76 imm <<= 4; 77 rol += 2; 78 } 79 80 if ((imm & 0xc0000000) == 0) { 81 imm <<= 2; 82 rol += 1; 83 } 84 85 if ((imm & 0x00ffffff) == 0) 86 return OP2_IMM | (imm >> 24) | (rol << 8); 87 88 return INVALID_IMM; 89 } 90 91 int ARMAssembler::genInt(int reg, ARMWord imm, bool positive) 92 { 93 // Step1: Search a non-immediate part 94 ARMWord mask; 95 ARMWord imm1; 96 ARMWord imm2; 97 int rol; 98 99 mask = 0xff000000; 100 rol = 8; 101 while(1) { 102 if ((imm & mask) == 0) { 103 imm = (imm << rol) | (imm >> (32 - rol)); 104 rol = 4 + (rol >> 1); 105 break; 106 } 107 rol += 2; 108 mask >>= 2; 109 if (mask & 0x3) { 110 // rol 8 111 imm = (imm << 8) | (imm >> 24); 112 mask = 0xff00; 113 rol = 24; 114 while (1) { 115 if ((imm & mask) == 0) { 116 imm = (imm << rol) | (imm >> (32 - rol)); 117 rol = (rol >> 1) - 8; 118 break; 119 } 120 rol += 2; 121 mask >>= 2; 122 if (mask & 0x3) 123 return 0; 124 } 125 break; 126 } 127 } 128 129 ASSERT((imm & 0xff) == 0); 130 131 if ((imm & 0xff000000) == 0) { 132 imm1 = OP2_IMM | ((imm >> 16) & 0xff) | (((rol + 4) & 0xf) << 8); 133 imm2 = OP2_IMM | ((imm >> 8) & 0xff) | (((rol + 8) & 0xf) << 8); 134 } else if (imm & 0xc0000000) { 135 imm1 = OP2_IMM | ((imm >> 24) & 0xff) | ((rol & 0xf) << 8); 136 imm <<= 8; 137 rol += 4; 138 139 if ((imm & 0xff000000) == 0) { 140 imm <<= 8; 141 rol += 4; 142 } 143 144 if ((imm & 0xf0000000) == 0) { 145 imm <<= 4; 146 rol += 2; 147 } 148 149 if ((imm & 0xc0000000) == 0) { 150 imm <<= 2; 151 rol += 1; 152 } 153 154 if ((imm & 0x00ffffff) == 0) 155 imm2 = OP2_IMM | (imm >> 24) | ((rol & 0xf) << 8); 156 else 157 return 0; 158 } else { 159 if ((imm & 0xf0000000) == 0) { 160 imm <<= 4; 161 rol += 2; 162 } 163 164 if ((imm & 0xc0000000) == 0) { 165 imm <<= 2; 166 rol += 1; 167 } 168 169 imm1 = OP2_IMM | ((imm >> 24) & 0xff) | ((rol & 0xf) << 8); 170 imm <<= 8; 171 rol += 4; 172 173 if ((imm & 0xf0000000) == 0) { 174 imm <<= 4; 175 rol += 2; 176 } 177 178 if ((imm & 0xc0000000) == 0) { 179 imm <<= 2; 180 rol += 1; 181 } 182 183 if ((imm & 0x00ffffff) == 0) 184 imm2 = OP2_IMM | (imm >> 24) | ((rol & 0xf) << 8); 185 else 186 return 0; 187 } 188 189 if (positive) { 190 mov_r(reg, imm1); 191 orr_r(reg, reg, imm2); 192 } else { 193 mvn_r(reg, imm1); 194 bic_r(reg, reg, imm2); 195 } 196 197 return 1; 198 } 199 200 ARMWord ARMAssembler::getImm(ARMWord imm, int tmpReg, bool invert) 201 { 202 ARMWord tmp; 203 204 // Do it by 1 instruction 205 tmp = getOp2(imm); 206 if (tmp != INVALID_IMM) 207 return tmp; 208 209 tmp = getOp2(~imm); 210 if (tmp != INVALID_IMM) { 211 if (invert) 212 return tmp | OP2_INV_IMM; 213 mvn_r(tmpReg, tmp); 214 return tmpReg; 215 } 216 217 return encodeComplexImm(imm, tmpReg); 218 } 219 220 void ARMAssembler::moveImm(ARMWord imm, int dest) 221 { 222 ARMWord tmp; 223 224 // Do it by 1 instruction 225 tmp = getOp2(imm); 226 if (tmp != INVALID_IMM) { 227 mov_r(dest, tmp); 228 return; 229 } 230 231 tmp = getOp2(~imm); 232 if (tmp != INVALID_IMM) { 233 mvn_r(dest, tmp); 234 return; 235 } 236 237 encodeComplexImm(imm, dest); 238 } 239 240 ARMWord ARMAssembler::encodeComplexImm(ARMWord imm, int dest) 241 { 242 #if WTF_ARM_ARCH_AT_LEAST(7) 243 ARMWord tmp = getImm16Op2(imm); 244 if (tmp != INVALID_IMM) { 245 movw_r(dest, tmp); 246 return dest; 247 } 248 movw_r(dest, getImm16Op2(imm & 0xffff)); 249 movt_r(dest, getImm16Op2(imm >> 16)); 250 return dest; 251 #else 252 // Do it by 2 instruction 253 if (genInt(dest, imm, true)) 254 return dest; 255 if (genInt(dest, ~imm, false)) 256 return dest; 257 258 ldr_imm(dest, imm); 259 return dest; 260 #endif 261 } 262 263 // Memory load/store helpers 264 265 void ARMAssembler::dataTransfer32(bool isLoad, RegisterID srcDst, RegisterID base, int32_t offset, bool bytes) 266 { 267 ARMWord transferFlag = bytes ? DT_BYTE : 0; 268 if (offset >= 0) { 269 if (offset <= 0xfff) 270 dtr_u(isLoad, srcDst, base, offset | transferFlag); 271 else if (offset <= 0xfffff) { 272 add_r(ARMRegisters::S0, base, OP2_IMM | (offset >> 12) | (10 << 8)); 273 dtr_u(isLoad, srcDst, ARMRegisters::S0, (offset & 0xfff) | transferFlag); 274 } else { 275 moveImm(offset, ARMRegisters::S0); 276 dtr_ur(isLoad, srcDst, base, ARMRegisters::S0 | transferFlag); 277 } 278 } else { 279 offset = -offset; 280 if (offset <= 0xfff) 281 dtr_d(isLoad, srcDst, base, offset | transferFlag); 282 else if (offset <= 0xfffff) { 283 sub_r(ARMRegisters::S0, base, OP2_IMM | (offset >> 12) | (10 << 8)); 284 dtr_d(isLoad, srcDst, ARMRegisters::S0, (offset & 0xfff) | transferFlag); 285 } else { 286 moveImm(offset, ARMRegisters::S0); 287 dtr_dr(isLoad, srcDst, base, ARMRegisters::S0 | transferFlag); 288 } 289 } 290 } 291 292 void ARMAssembler::baseIndexTransfer32(bool isLoad, RegisterID srcDst, RegisterID base, RegisterID index, int scale, int32_t offset) 293 { 294 ARMWord op2; 295 296 ASSERT(scale >= 0 && scale <= 3); 297 op2 = lsl(index, scale); 298 299 if (offset >= 0 && offset <= 0xfff) { 300 add_r(ARMRegisters::S0, base, op2); 301 dtr_u(isLoad, srcDst, ARMRegisters::S0, offset); 302 return; 303 } 304 if (offset <= 0 && offset >= -0xfff) { 305 add_r(ARMRegisters::S0, base, op2); 306 dtr_d(isLoad, srcDst, ARMRegisters::S0, -offset); 307 return; 308 } 309 310 ldr_un_imm(ARMRegisters::S0, offset); 311 add_r(ARMRegisters::S0, ARMRegisters::S0, op2); 312 dtr_ur(isLoad, srcDst, base, ARMRegisters::S0); 313 } 314 315 void ARMAssembler::doubleTransfer(bool isLoad, FPRegisterID srcDst, RegisterID base, int32_t offset) 316 { 317 if (offset & 0x3) { 318 if (offset <= 0x3ff && offset >= 0) { 319 fdtr_u(isLoad, srcDst, base, offset >> 2); 320 return; 321 } 322 if (offset <= 0x3ffff && offset >= 0) { 323 add_r(ARMRegisters::S0, base, OP2_IMM | (offset >> 10) | (11 << 8)); 324 fdtr_u(isLoad, srcDst, ARMRegisters::S0, (offset >> 2) & 0xff); 325 return; 326 } 327 offset = -offset; 328 329 if (offset <= 0x3ff && offset >= 0) { 330 fdtr_d(isLoad, srcDst, base, offset >> 2); 331 return; 332 } 333 if (offset <= 0x3ffff && offset >= 0) { 334 sub_r(ARMRegisters::S0, base, OP2_IMM | (offset >> 10) | (11 << 8)); 335 fdtr_d(isLoad, srcDst, ARMRegisters::S0, (offset >> 2) & 0xff); 336 return; 337 } 338 offset = -offset; 339 } 340 341 ldr_un_imm(ARMRegisters::S0, offset); 342 add_r(ARMRegisters::S0, ARMRegisters::S0, base); 343 fdtr_u(isLoad, srcDst, ARMRegisters::S0, 0); 344 } 345 346 void* ARMAssembler::executableCopy(ExecutablePool* allocator) 347 { 348 // 64-bit alignment is required for next constant pool and JIT code as well 349 m_buffer.flushWithoutBarrier(true); 350 if (m_buffer.uncheckedSize() & 0x7) 351 bkpt(0); 352 353 char* data = reinterpret_cast<char*>(m_buffer.executableCopy(allocator)); 354 355 for (Jumps::Iterator iter = m_jumps.begin(); iter != m_jumps.end(); ++iter) { 356 // The last bit is set if the constant must be placed on constant pool. 357 int pos = (*iter) & (~0x1); 358 ARMWord* ldrAddr = reinterpret_cast_ptr<ARMWord*>(data + pos); 359 ARMWord* addr = getLdrImmAddress(ldrAddr); 360 if (*addr != InvalidBranchTarget) { 361 if (!(*iter & 1)) { 362 int diff = reinterpret_cast_ptr<ARMWord*>(data + *addr) - (ldrAddr + DefaultPrefetching); 363 364 if ((diff <= BOFFSET_MAX && diff >= BOFFSET_MIN)) { 365 *ldrAddr = B | getConditionalField(*ldrAddr) | (diff & BRANCH_MASK); 366 continue; 367 } 368 } 369 *addr = reinterpret_cast<ARMWord>(data + *addr); 370 } 371 } 372 373 return data; 374 } 375 376 } // namespace JSC 377 378 #endif // ENABLE(ASSEMBLER) && CPU(ARM_TRADITIONAL) 379