1 /* 2 * Stack-less Just-In-Time compiler 3 * 4 * Copyright Zoltan Herczeg (hzmester (at) freemail.hu). All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without modification, are 7 * permitted provided that the following conditions are met: 8 * 9 * 1. Redistributions of source code must retain the above copyright notice, this list of 10 * conditions and the following disclaimer. 11 * 12 * 2. Redistributions in binary form must reproduce the above copyright notice, this list 13 * of conditions and the following disclaimer in the documentation and/or other materials 14 * provided with the distribution. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY 17 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT 19 * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED 21 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR 22 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 23 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN 24 * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 25 */ 26 27 SLJIT_API_FUNC_ATTRIBUTE const char* sljit_get_platform_name(void) 28 { 29 return "ARM-64" SLJIT_CPUINFO; 30 } 31 32 /* Length of an instruction word */ 33 typedef sljit_u32 sljit_ins; 34 35 #define TMP_ZERO (0) 36 37 #define TMP_REG1 (SLJIT_NUMBER_OF_REGISTERS + 2) 38 #define TMP_REG2 (SLJIT_NUMBER_OF_REGISTERS + 3) 39 #define TMP_LR (SLJIT_NUMBER_OF_REGISTERS + 4) 40 #define TMP_FP (SLJIT_NUMBER_OF_REGISTERS + 5) 41 42 #define TMP_FREG1 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1) 43 #define TMP_FREG2 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 2) 44 45 /* r18 - platform register, currently not used */ 46 static const sljit_u8 reg_map[SLJIT_NUMBER_OF_REGISTERS + 8] = { 47 31, 0, 1, 2, 3, 4, 5, 6, 7, 11, 12, 13, 14, 15, 16, 17, 8, 28, 27, 26, 25, 24, 23, 22, 21, 20, 19, 31, 9, 10, 30, 29 48 }; 49 50 static const sljit_u8 freg_map[SLJIT_NUMBER_OF_FLOAT_REGISTERS + 3] = { 51 0, 0, 1, 2, 3, 4, 5, 6, 7 52 }; 53 54 #define W_OP (1 << 31) 55 #define RD(rd) (reg_map[rd]) 56 #define RT(rt) (reg_map[rt]) 57 #define RN(rn) (reg_map[rn] << 5) 58 #define RT2(rt2) (reg_map[rt2] << 10) 59 #define RM(rm) (reg_map[rm] << 16) 60 #define VD(vd) (freg_map[vd]) 61 #define VT(vt) (freg_map[vt]) 62 #define VN(vn) (freg_map[vn] << 5) 63 #define VM(vm) (freg_map[vm] << 16) 64 65 /* --------------------------------------------------------------------- */ 66 /* Instrucion forms */ 67 /* --------------------------------------------------------------------- */ 68 69 #define ADC 0x9a000000 70 #define ADD 0x8b000000 71 #define ADDE 0x8b200000 72 #define ADDI 0x91000000 73 #define AND 0x8a000000 74 #define ANDI 0x92000000 75 #define ASRV 0x9ac02800 76 #define B 0x14000000 77 #define B_CC 0x54000000 78 #define BL 0x94000000 79 #define BLR 0xd63f0000 80 #define BR 0xd61f0000 81 #define BRK 0xd4200000 82 #define CBZ 0xb4000000 83 #define CLZ 0xdac01000 84 #define CSEL 0x9a800000 85 #define CSINC 0x9a800400 86 #define EOR 0xca000000 87 #define EORI 0xd2000000 88 #define FABS 0x1e60c000 89 #define FADD 0x1e602800 90 #define FCMP 0x1e602000 91 #define FCVT 0x1e224000 92 #define FCVTZS 0x9e780000 93 #define FDIV 0x1e601800 94 #define FMOV 0x1e604000 95 #define FMUL 0x1e600800 96 #define FNEG 0x1e614000 97 #define FSUB 0x1e603800 98 #define LDRI 0xf9400000 99 #define LDP 0xa9400000 100 #define LDP_PRE 0xa9c00000 101 #define LDR_PRE 0xf8400c00 102 #define LSLV 0x9ac02000 103 #define LSRV 0x9ac02400 104 #define MADD 0x9b000000 105 #define MOVK 0xf2800000 106 #define MOVN 0x92800000 107 #define MOVZ 0xd2800000 108 #define NOP 0xd503201f 109 #define ORN 0xaa200000 110 #define ORR 0xaa000000 111 #define ORRI 0xb2000000 112 #define RET 0xd65f0000 113 #define SBC 0xda000000 114 #define SBFM 0x93000000 115 #define SCVTF 0x9e620000 116 #define SDIV 0x9ac00c00 117 #define SMADDL 0x9b200000 118 #define SMULH 0x9b403c00 119 #define STP 0xa9000000 120 #define STP_PRE 0xa9800000 121 #define STRB 0x38206800 122 #define STRBI 0x39000000 123 #define STRI 0xf9000000 124 #define STR_FI 0x3d000000 125 #define STR_FR 0x3c206800 126 #define STUR_FI 0x3c000000 127 #define STURBI 0x38000000 128 #define SUB 0xcb000000 129 #define SUBI 0xd1000000 130 #define SUBS 0xeb000000 131 #define UBFM 0xd3000000 132 #define UDIV 0x9ac00800 133 #define UMULH 0x9bc03c00 134 135 /* dest_reg is the absolute name of the register 136 Useful for reordering instructions in the delay slot. */ 137 static sljit_s32 push_inst(struct sljit_compiler *compiler, sljit_ins ins) 138 { 139 sljit_ins *ptr = (sljit_ins*)ensure_buf(compiler, sizeof(sljit_ins)); 140 FAIL_IF(!ptr); 141 *ptr = ins; 142 compiler->size++; 143 return SLJIT_SUCCESS; 144 } 145 146 static SLJIT_INLINE sljit_s32 emit_imm64_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_uw imm) 147 { 148 FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((imm & 0xffff) << 5))); 149 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | (((imm >> 16) & 0xffff) << 5) | (1 << 21))); 150 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | (((imm >> 32) & 0xffff) << 5) | (2 << 21))); 151 return push_inst(compiler, MOVK | RD(dst) | ((imm >> 48) << 5) | (3 << 21)); 152 } 153 154 static SLJIT_INLINE void modify_imm64_const(sljit_ins* inst, sljit_uw new_imm) 155 { 156 sljit_s32 dst = inst[0] & 0x1f; 157 SLJIT_ASSERT((inst[0] & 0xffe00000) == MOVZ && (inst[1] & 0xffe00000) == (MOVK | (1 << 21))); 158 inst[0] = MOVZ | dst | ((new_imm & 0xffff) << 5); 159 inst[1] = MOVK | dst | (((new_imm >> 16) & 0xffff) << 5) | (1 << 21); 160 inst[2] = MOVK | dst | (((new_imm >> 32) & 0xffff) << 5) | (2 << 21); 161 inst[3] = MOVK | dst | ((new_imm >> 48) << 5) | (3 << 21); 162 } 163 164 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset) 165 { 166 sljit_sw diff; 167 sljit_uw target_addr; 168 169 if (jump->flags & SLJIT_REWRITABLE_JUMP) { 170 jump->flags |= PATCH_ABS64; 171 return 0; 172 } 173 174 if (jump->flags & JUMP_ADDR) 175 target_addr = jump->u.target; 176 else { 177 SLJIT_ASSERT(jump->flags & JUMP_LABEL); 178 target_addr = (sljit_uw)(code + jump->u.label->size) + (sljit_uw)executable_offset; 179 } 180 181 diff = (sljit_sw)target_addr - (sljit_sw)(code_ptr + 4) - executable_offset; 182 183 if (jump->flags & IS_COND) { 184 diff += sizeof(sljit_ins); 185 if (diff <= 0xfffff && diff >= -0x100000) { 186 code_ptr[-5] ^= (jump->flags & IS_CBZ) ? (0x1 << 24) : 0x1; 187 jump->addr -= sizeof(sljit_ins); 188 jump->flags |= PATCH_COND; 189 return 5; 190 } 191 diff -= sizeof(sljit_ins); 192 } 193 194 if (diff <= 0x7ffffff && diff >= -0x8000000) { 195 jump->flags |= PATCH_B; 196 return 4; 197 } 198 199 if (target_addr <= 0xffffffffl) { 200 if (jump->flags & IS_COND) 201 code_ptr[-5] -= (2 << 5); 202 code_ptr[-2] = code_ptr[0]; 203 return 2; 204 } 205 206 if (target_addr <= 0xffffffffffffl) { 207 if (jump->flags & IS_COND) 208 code_ptr[-5] -= (1 << 5); 209 jump->flags |= PATCH_ABS48; 210 code_ptr[-1] = code_ptr[0]; 211 return 1; 212 } 213 214 jump->flags |= PATCH_ABS64; 215 return 0; 216 } 217 218 SLJIT_API_FUNC_ATTRIBUTE void* sljit_generate_code(struct sljit_compiler *compiler) 219 { 220 struct sljit_memory_fragment *buf; 221 sljit_ins *code; 222 sljit_ins *code_ptr; 223 sljit_ins *buf_ptr; 224 sljit_ins *buf_end; 225 sljit_uw word_count; 226 sljit_sw executable_offset; 227 sljit_uw addr; 228 sljit_s32 dst; 229 230 struct sljit_label *label; 231 struct sljit_jump *jump; 232 struct sljit_const *const_; 233 234 CHECK_ERROR_PTR(); 235 CHECK_PTR(check_sljit_generate_code(compiler)); 236 reverse_buf(compiler); 237 238 code = (sljit_ins*)SLJIT_MALLOC_EXEC(compiler->size * sizeof(sljit_ins)); 239 PTR_FAIL_WITH_EXEC_IF(code); 240 buf = compiler->buf; 241 242 code_ptr = code; 243 word_count = 0; 244 executable_offset = SLJIT_EXEC_OFFSET(code); 245 246 label = compiler->labels; 247 jump = compiler->jumps; 248 const_ = compiler->consts; 249 250 do { 251 buf_ptr = (sljit_ins*)buf->memory; 252 buf_end = buf_ptr + (buf->used_size >> 2); 253 do { 254 *code_ptr = *buf_ptr++; 255 /* These structures are ordered by their address. */ 256 SLJIT_ASSERT(!label || label->size >= word_count); 257 SLJIT_ASSERT(!jump || jump->addr >= word_count); 258 SLJIT_ASSERT(!const_ || const_->addr >= word_count); 259 if (label && label->size == word_count) { 260 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 261 label->size = code_ptr - code; 262 label = label->next; 263 } 264 if (jump && jump->addr == word_count) { 265 jump->addr = (sljit_uw)(code_ptr - 4); 266 code_ptr -= detect_jump_type(jump, code_ptr, code, executable_offset); 267 jump = jump->next; 268 } 269 if (const_ && const_->addr == word_count) { 270 const_->addr = (sljit_uw)code_ptr; 271 const_ = const_->next; 272 } 273 code_ptr ++; 274 word_count ++; 275 } while (buf_ptr < buf_end); 276 277 buf = buf->next; 278 } while (buf); 279 280 if (label && label->size == word_count) { 281 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 282 label->size = code_ptr - code; 283 label = label->next; 284 } 285 286 SLJIT_ASSERT(!label); 287 SLJIT_ASSERT(!jump); 288 SLJIT_ASSERT(!const_); 289 SLJIT_ASSERT(code_ptr - code <= (sljit_sw)compiler->size); 290 291 jump = compiler->jumps; 292 while (jump) { 293 do { 294 addr = (jump->flags & JUMP_LABEL) ? jump->u.label->addr : jump->u.target; 295 buf_ptr = (sljit_ins *)jump->addr; 296 297 if (jump->flags & PATCH_B) { 298 addr = (sljit_sw)(addr - (sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset)) >> 2; 299 SLJIT_ASSERT((sljit_sw)addr <= 0x1ffffff && (sljit_sw)addr >= -0x2000000); 300 buf_ptr[0] = ((jump->flags & IS_BL) ? BL : B) | (addr & 0x3ffffff); 301 if (jump->flags & IS_COND) 302 buf_ptr[-1] -= (4 << 5); 303 break; 304 } 305 if (jump->flags & PATCH_COND) { 306 addr = (sljit_sw)(addr - (sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset)) >> 2; 307 SLJIT_ASSERT((sljit_sw)addr <= 0x3ffff && (sljit_sw)addr >= -0x40000); 308 buf_ptr[0] = (buf_ptr[0] & ~0xffffe0) | ((addr & 0x7ffff) << 5); 309 break; 310 } 311 312 SLJIT_ASSERT((jump->flags & (PATCH_ABS48 | PATCH_ABS64)) || addr <= 0xffffffffl); 313 SLJIT_ASSERT((jump->flags & PATCH_ABS64) || addr <= 0xffffffffffffl); 314 315 dst = buf_ptr[0] & 0x1f; 316 buf_ptr[0] = MOVZ | dst | ((addr & 0xffff) << 5); 317 buf_ptr[1] = MOVK | dst | (((addr >> 16) & 0xffff) << 5) | (1 << 21); 318 if (jump->flags & (PATCH_ABS48 | PATCH_ABS64)) 319 buf_ptr[2] = MOVK | dst | (((addr >> 32) & 0xffff) << 5) | (2 << 21); 320 if (jump->flags & PATCH_ABS64) 321 buf_ptr[3] = MOVK | dst | (((addr >> 48) & 0xffff) << 5) | (3 << 21); 322 } while (0); 323 jump = jump->next; 324 } 325 326 compiler->error = SLJIT_ERR_COMPILED; 327 compiler->executable_offset = executable_offset; 328 compiler->executable_size = (code_ptr - code) * sizeof(sljit_ins); 329 330 code = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code, executable_offset); 331 code_ptr = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 332 333 SLJIT_CACHE_FLUSH(code, code_ptr); 334 return code; 335 } 336 337 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_has_cpu_feature(sljit_s32 feature_type) 338 { 339 switch (feature_type) { 340 case SLJIT_HAS_FPU: 341 #ifdef SLJIT_IS_FPU_AVAILABLE 342 return SLJIT_IS_FPU_AVAILABLE; 343 #else 344 /* Available by default. */ 345 return 1; 346 #endif 347 348 case SLJIT_HAS_CLZ: 349 case SLJIT_HAS_CMOV: 350 return 1; 351 352 default: 353 return 0; 354 } 355 } 356 357 /* --------------------------------------------------------------------- */ 358 /* Core code generator functions. */ 359 /* --------------------------------------------------------------------- */ 360 361 #define COUNT_TRAILING_ZERO(value, result) \ 362 result = 0; \ 363 if (!(value & 0xffffffff)) { \ 364 result += 32; \ 365 value >>= 32; \ 366 } \ 367 if (!(value & 0xffff)) { \ 368 result += 16; \ 369 value >>= 16; \ 370 } \ 371 if (!(value & 0xff)) { \ 372 result += 8; \ 373 value >>= 8; \ 374 } \ 375 if (!(value & 0xf)) { \ 376 result += 4; \ 377 value >>= 4; \ 378 } \ 379 if (!(value & 0x3)) { \ 380 result += 2; \ 381 value >>= 2; \ 382 } \ 383 if (!(value & 0x1)) { \ 384 result += 1; \ 385 value >>= 1; \ 386 } 387 388 #define LOGICAL_IMM_CHECK 0x100 389 390 static sljit_ins logical_imm(sljit_sw imm, sljit_s32 len) 391 { 392 sljit_s32 negated, ones, right; 393 sljit_uw mask, uimm; 394 sljit_ins ins; 395 396 if (len & LOGICAL_IMM_CHECK) { 397 len &= ~LOGICAL_IMM_CHECK; 398 if (len == 32 && (imm == 0 || imm == -1)) 399 return 0; 400 if (len == 16 && ((sljit_s32)imm == 0 || (sljit_s32)imm == -1)) 401 return 0; 402 } 403 404 SLJIT_ASSERT((len == 32 && imm != 0 && imm != -1) 405 || (len == 16 && (sljit_s32)imm != 0 && (sljit_s32)imm != -1)); 406 407 uimm = (sljit_uw)imm; 408 while (1) { 409 if (len <= 0) { 410 SLJIT_UNREACHABLE(); 411 return 0; 412 } 413 414 mask = ((sljit_uw)1 << len) - 1; 415 if ((uimm & mask) != ((uimm >> len) & mask)) 416 break; 417 len >>= 1; 418 } 419 420 len <<= 1; 421 422 negated = 0; 423 if (uimm & 0x1) { 424 negated = 1; 425 uimm = ~uimm; 426 } 427 428 if (len < 64) 429 uimm &= ((sljit_uw)1 << len) - 1; 430 431 /* Unsigned right shift. */ 432 COUNT_TRAILING_ZERO(uimm, right); 433 434 /* Signed shift. We also know that the highest bit is set. */ 435 imm = (sljit_sw)~uimm; 436 SLJIT_ASSERT(imm < 0); 437 438 COUNT_TRAILING_ZERO(imm, ones); 439 440 if (~imm) 441 return 0; 442 443 if (len == 64) 444 ins = 1 << 22; 445 else 446 ins = (0x3f - ((len << 1) - 1)) << 10; 447 448 if (negated) 449 return ins | ((len - ones - 1) << 10) | ((len - ones - right) << 16); 450 451 return ins | ((ones - 1) << 10) | ((len - right) << 16); 452 } 453 454 #undef COUNT_TRAILING_ZERO 455 456 static sljit_s32 load_immediate(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw simm) 457 { 458 sljit_uw imm = (sljit_uw)simm; 459 sljit_s32 i, zeros, ones, first; 460 sljit_ins bitmask; 461 462 /* Handling simple immediates first. */ 463 if (imm <= 0xffff) 464 return push_inst(compiler, MOVZ | RD(dst) | (imm << 5)); 465 466 if (simm < 0 && simm >= -0x10000) 467 return push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff) << 5)); 468 469 if (imm <= 0xffffffffl) { 470 if ((imm & 0xffff) == 0) 471 return push_inst(compiler, MOVZ | RD(dst) | ((imm >> 16) << 5) | (1 << 21)); 472 if ((imm & 0xffff0000l) == 0xffff0000) 473 return push_inst(compiler, (MOVN ^ W_OP) | RD(dst) | ((~imm & 0xffff) << 5)); 474 if ((imm & 0xffff) == 0xffff) 475 return push_inst(compiler, (MOVN ^ W_OP) | RD(dst) | ((~imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 476 477 bitmask = logical_imm(simm, 16); 478 if (bitmask != 0) 479 return push_inst(compiler, (ORRI ^ W_OP) | RD(dst) | RN(TMP_ZERO) | bitmask); 480 481 FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((imm & 0xffff) << 5))); 482 return push_inst(compiler, MOVK | RD(dst) | ((imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 483 } 484 485 bitmask = logical_imm(simm, 32); 486 if (bitmask != 0) 487 return push_inst(compiler, ORRI | RD(dst) | RN(TMP_ZERO) | bitmask); 488 489 if (simm < 0 && simm >= -0x100000000l) { 490 if ((imm & 0xffff) == 0xffff) 491 return push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 492 493 FAIL_IF(push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff) << 5))); 494 return push_inst(compiler, MOVK | RD(dst) | ((imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 495 } 496 497 /* A large amount of number can be constructed from ORR and MOVx, but computing them is costly. */ 498 499 zeros = 0; 500 ones = 0; 501 for (i = 4; i > 0; i--) { 502 if ((simm & 0xffff) == 0) 503 zeros++; 504 if ((simm & 0xffff) == 0xffff) 505 ones++; 506 simm >>= 16; 507 } 508 509 simm = (sljit_sw)imm; 510 first = 1; 511 if (ones > zeros) { 512 simm = ~simm; 513 for (i = 0; i < 4; i++) { 514 if (!(simm & 0xffff)) { 515 simm >>= 16; 516 continue; 517 } 518 if (first) { 519 first = 0; 520 FAIL_IF(push_inst(compiler, MOVN | RD(dst) | ((simm & 0xffff) << 5) | (i << 21))); 521 } 522 else 523 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | ((~simm & 0xffff) << 5) | (i << 21))); 524 simm >>= 16; 525 } 526 return SLJIT_SUCCESS; 527 } 528 529 for (i = 0; i < 4; i++) { 530 if (!(simm & 0xffff)) { 531 simm >>= 16; 532 continue; 533 } 534 if (first) { 535 first = 0; 536 FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((simm & 0xffff) << 5) | (i << 21))); 537 } 538 else 539 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | ((simm & 0xffff) << 5) | (i << 21))); 540 simm >>= 16; 541 } 542 return SLJIT_SUCCESS; 543 } 544 545 #define ARG1_IMM 0x0010000 546 #define ARG2_IMM 0x0020000 547 #define INT_OP 0x0040000 548 #define SET_FLAGS 0x0080000 549 #define UNUSED_RETURN 0x0100000 550 551 #define CHECK_FLAGS(flag_bits) \ 552 if (flags & SET_FLAGS) { \ 553 inv_bits |= flag_bits; \ 554 if (flags & UNUSED_RETURN) \ 555 dst = TMP_ZERO; \ 556 } 557 558 static sljit_s32 emit_op_imm(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 dst, sljit_sw arg1, sljit_sw arg2) 559 { 560 /* dst must be register, TMP_REG1 561 arg1 must be register, TMP_REG1, imm 562 arg2 must be register, TMP_REG2, imm */ 563 sljit_ins inv_bits = (flags & INT_OP) ? (1 << 31) : 0; 564 sljit_ins inst_bits; 565 sljit_s32 op = (flags & 0xffff); 566 sljit_s32 reg; 567 sljit_sw imm, nimm; 568 569 if (SLJIT_UNLIKELY((flags & (ARG1_IMM | ARG2_IMM)) == (ARG1_IMM | ARG2_IMM))) { 570 /* Both are immediates. */ 571 flags &= ~ARG1_IMM; 572 if (arg1 == 0 && op != SLJIT_ADD && op != SLJIT_SUB) 573 arg1 = TMP_ZERO; 574 else { 575 FAIL_IF(load_immediate(compiler, TMP_REG1, arg1)); 576 arg1 = TMP_REG1; 577 } 578 } 579 580 if (flags & (ARG1_IMM | ARG2_IMM)) { 581 reg = (flags & ARG2_IMM) ? arg1 : arg2; 582 imm = (flags & ARG2_IMM) ? arg2 : arg1; 583 584 switch (op) { 585 case SLJIT_MUL: 586 case SLJIT_NEG: 587 case SLJIT_CLZ: 588 case SLJIT_ADDC: 589 case SLJIT_SUBC: 590 /* No form with immediate operand (except imm 0, which 591 is represented by a ZERO register). */ 592 break; 593 case SLJIT_MOV: 594 SLJIT_ASSERT(!(flags & SET_FLAGS) && (flags & ARG2_IMM) && arg1 == TMP_REG1); 595 return load_immediate(compiler, dst, imm); 596 case SLJIT_NOT: 597 SLJIT_ASSERT(flags & ARG2_IMM); 598 FAIL_IF(load_immediate(compiler, dst, (flags & INT_OP) ? (~imm & 0xffffffff) : ~imm)); 599 goto set_flags; 600 case SLJIT_SUB: 601 if (flags & ARG1_IMM) 602 break; 603 imm = -imm; 604 /* Fall through. */ 605 case SLJIT_ADD: 606 if (imm == 0) { 607 CHECK_FLAGS(1 << 29); 608 return push_inst(compiler, ((op == SLJIT_ADD ? ADDI : SUBI) ^ inv_bits) | RD(dst) | RN(reg)); 609 } 610 if (imm > 0 && imm <= 0xfff) { 611 CHECK_FLAGS(1 << 29); 612 return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | (imm << 10)); 613 } 614 nimm = -imm; 615 if (nimm > 0 && nimm <= 0xfff) { 616 CHECK_FLAGS(1 << 29); 617 return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | (nimm << 10)); 618 } 619 if (imm > 0 && imm <= 0xffffff && !(imm & 0xfff)) { 620 CHECK_FLAGS(1 << 29); 621 return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | ((imm >> 12) << 10) | (1 << 22)); 622 } 623 if (nimm > 0 && nimm <= 0xffffff && !(nimm & 0xfff)) { 624 CHECK_FLAGS(1 << 29); 625 return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | ((nimm >> 12) << 10) | (1 << 22)); 626 } 627 if (imm > 0 && imm <= 0xffffff && !(flags & SET_FLAGS)) { 628 FAIL_IF(push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | ((imm >> 12) << 10) | (1 << 22))); 629 return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(dst) | ((imm & 0xfff) << 10)); 630 } 631 if (nimm > 0 && nimm <= 0xffffff && !(flags & SET_FLAGS)) { 632 FAIL_IF(push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | ((nimm >> 12) << 10) | (1 << 22))); 633 return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(dst) | ((nimm & 0xfff) << 10)); 634 } 635 break; 636 case SLJIT_AND: 637 inst_bits = logical_imm(imm, LOGICAL_IMM_CHECK | ((flags & INT_OP) ? 16 : 32)); 638 if (!inst_bits) 639 break; 640 CHECK_FLAGS(3 << 29); 641 return push_inst(compiler, (ANDI ^ inv_bits) | RD(dst) | RN(reg) | inst_bits); 642 case SLJIT_OR: 643 case SLJIT_XOR: 644 inst_bits = logical_imm(imm, LOGICAL_IMM_CHECK | ((flags & INT_OP) ? 16 : 32)); 645 if (!inst_bits) 646 break; 647 if (op == SLJIT_OR) 648 inst_bits |= ORRI; 649 else 650 inst_bits |= EORI; 651 FAIL_IF(push_inst(compiler, (inst_bits ^ inv_bits) | RD(dst) | RN(reg))); 652 goto set_flags; 653 case SLJIT_SHL: 654 if (flags & ARG1_IMM) 655 break; 656 if (flags & INT_OP) { 657 imm &= 0x1f; 658 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | ((-imm & 0x1f) << 16) | ((31 - imm) << 10))); 659 } 660 else { 661 imm &= 0x3f; 662 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (1 << 22) | ((-imm & 0x3f) << 16) | ((63 - imm) << 10))); 663 } 664 goto set_flags; 665 case SLJIT_LSHR: 666 case SLJIT_ASHR: 667 if (flags & ARG1_IMM) 668 break; 669 if (op == SLJIT_ASHR) 670 inv_bits |= 1 << 30; 671 if (flags & INT_OP) { 672 imm &= 0x1f; 673 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (imm << 16) | (31 << 10))); 674 } 675 else { 676 imm &= 0x3f; 677 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (1 << 22) | (imm << 16) | (63 << 10))); 678 } 679 goto set_flags; 680 default: 681 SLJIT_UNREACHABLE(); 682 break; 683 } 684 685 if (flags & ARG2_IMM) { 686 if (arg2 == 0) 687 arg2 = TMP_ZERO; 688 else { 689 FAIL_IF(load_immediate(compiler, TMP_REG2, arg2)); 690 arg2 = TMP_REG2; 691 } 692 } 693 else { 694 if (arg1 == 0) 695 arg1 = TMP_ZERO; 696 else { 697 FAIL_IF(load_immediate(compiler, TMP_REG1, arg1)); 698 arg1 = TMP_REG1; 699 } 700 } 701 } 702 703 /* Both arguments are registers. */ 704 switch (op) { 705 case SLJIT_MOV: 706 case SLJIT_MOV_P: 707 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 708 if (dst == arg2) 709 return SLJIT_SUCCESS; 710 return push_inst(compiler, ORR | RD(dst) | RN(TMP_ZERO) | RM(arg2)); 711 case SLJIT_MOV_U8: 712 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 713 return push_inst(compiler, (UBFM ^ (1 << 31)) | RD(dst) | RN(arg2) | (7 << 10)); 714 case SLJIT_MOV_S8: 715 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 716 if (!(flags & INT_OP)) 717 inv_bits |= 1 << 22; 718 return push_inst(compiler, (SBFM ^ inv_bits) | RD(dst) | RN(arg2) | (7 << 10)); 719 case SLJIT_MOV_U16: 720 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 721 return push_inst(compiler, (UBFM ^ (1 << 31)) | RD(dst) | RN(arg2) | (15 << 10)); 722 case SLJIT_MOV_S16: 723 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 724 if (!(flags & INT_OP)) 725 inv_bits |= 1 << 22; 726 return push_inst(compiler, (SBFM ^ inv_bits) | RD(dst) | RN(arg2) | (15 << 10)); 727 case SLJIT_MOV_U32: 728 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 729 if ((flags & INT_OP) && dst == arg2) 730 return SLJIT_SUCCESS; 731 return push_inst(compiler, (ORR ^ (1 << 31)) | RD(dst) | RN(TMP_ZERO) | RM(arg2)); 732 case SLJIT_MOV_S32: 733 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 734 if ((flags & INT_OP) && dst == arg2) 735 return SLJIT_SUCCESS; 736 return push_inst(compiler, SBFM | (1 << 22) | RD(dst) | RN(arg2) | (31 << 10)); 737 case SLJIT_NOT: 738 SLJIT_ASSERT(arg1 == TMP_REG1); 739 FAIL_IF(push_inst(compiler, (ORN ^ inv_bits) | RD(dst) | RN(TMP_ZERO) | RM(arg2))); 740 break; /* Set flags. */ 741 case SLJIT_NEG: 742 SLJIT_ASSERT(arg1 == TMP_REG1); 743 if (flags & SET_FLAGS) 744 inv_bits |= 1 << 29; 745 return push_inst(compiler, (SUB ^ inv_bits) | RD(dst) | RN(TMP_ZERO) | RM(arg2)); 746 case SLJIT_CLZ: 747 SLJIT_ASSERT(arg1 == TMP_REG1); 748 return push_inst(compiler, (CLZ ^ inv_bits) | RD(dst) | RN(arg2)); 749 case SLJIT_ADD: 750 CHECK_FLAGS(1 << 29); 751 return push_inst(compiler, (ADD ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 752 case SLJIT_ADDC: 753 CHECK_FLAGS(1 << 29); 754 return push_inst(compiler, (ADC ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 755 case SLJIT_SUB: 756 CHECK_FLAGS(1 << 29); 757 return push_inst(compiler, (SUB ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 758 case SLJIT_SUBC: 759 CHECK_FLAGS(1 << 29); 760 return push_inst(compiler, (SBC ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 761 case SLJIT_MUL: 762 if (!(flags & SET_FLAGS)) 763 return push_inst(compiler, (MADD ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2) | RT2(TMP_ZERO)); 764 if (flags & INT_OP) { 765 FAIL_IF(push_inst(compiler, SMADDL | RD(dst) | RN(arg1) | RM(arg2) | (31 << 10))); 766 FAIL_IF(push_inst(compiler, ADD | RD(TMP_LR) | RN(TMP_ZERO) | RM(dst) | (2 << 22) | (31 << 10))); 767 return push_inst(compiler, SUBS | RD(TMP_ZERO) | RN(TMP_LR) | RM(dst) | (2 << 22) | (63 << 10)); 768 } 769 FAIL_IF(push_inst(compiler, SMULH | RD(TMP_LR) | RN(arg1) | RM(arg2))); 770 FAIL_IF(push_inst(compiler, MADD | RD(dst) | RN(arg1) | RM(arg2) | RT2(TMP_ZERO))); 771 return push_inst(compiler, SUBS | RD(TMP_ZERO) | RN(TMP_LR) | RM(dst) | (2 << 22) | (63 << 10)); 772 case SLJIT_AND: 773 CHECK_FLAGS(3 << 29); 774 return push_inst(compiler, (AND ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 775 case SLJIT_OR: 776 FAIL_IF(push_inst(compiler, (ORR ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 777 break; /* Set flags. */ 778 case SLJIT_XOR: 779 FAIL_IF(push_inst(compiler, (EOR ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 780 break; /* Set flags. */ 781 case SLJIT_SHL: 782 FAIL_IF(push_inst(compiler, (LSLV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 783 break; /* Set flags. */ 784 case SLJIT_LSHR: 785 FAIL_IF(push_inst(compiler, (LSRV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 786 break; /* Set flags. */ 787 case SLJIT_ASHR: 788 FAIL_IF(push_inst(compiler, (ASRV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 789 break; /* Set flags. */ 790 default: 791 SLJIT_UNREACHABLE(); 792 return SLJIT_SUCCESS; 793 } 794 795 set_flags: 796 if (flags & SET_FLAGS) 797 return push_inst(compiler, (SUBS ^ inv_bits) | RD(TMP_ZERO) | RN(dst) | RM(TMP_ZERO)); 798 return SLJIT_SUCCESS; 799 } 800 801 #define STORE 0x10 802 #define SIGNED 0x20 803 804 #define BYTE_SIZE 0x0 805 #define HALF_SIZE 0x1 806 #define INT_SIZE 0x2 807 #define WORD_SIZE 0x3 808 809 #define MEM_SIZE_SHIFT(flags) ((flags) & 0x3) 810 811 static sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, 812 sljit_s32 arg, sljit_sw argw, sljit_s32 tmp_reg) 813 { 814 sljit_u32 shift = MEM_SIZE_SHIFT(flags); 815 sljit_u32 type = (shift << 30); 816 817 if (!(flags & STORE)) 818 type |= (flags & SIGNED) ? 0x00800000 : 0x00400000; 819 820 SLJIT_ASSERT(arg & SLJIT_MEM); 821 822 if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) { 823 argw &= 0x3; 824 825 if (argw == 0 || argw == shift) 826 return push_inst(compiler, STRB | type | RT(reg) 827 | RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw ? (1 << 12) : 0)); 828 829 FAIL_IF(push_inst(compiler, ADD | RD(tmp_reg) | RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw << 10))); 830 return push_inst(compiler, STRBI | type | RT(reg) | RN(tmp_reg)); 831 } 832 833 arg &= REG_MASK; 834 835 if (arg == SLJIT_UNUSED) { 836 FAIL_IF(load_immediate(compiler, tmp_reg, argw & ~(0xfff << shift))); 837 838 argw = (argw >> shift) & 0xfff; 839 840 return push_inst(compiler, STRBI | type | RT(reg) | RN(tmp_reg) | (argw << 10)); 841 } 842 843 if (argw >= 0 && (argw & ((1 << shift) - 1)) == 0) { 844 if ((argw >> shift) <= 0xfff) { 845 return push_inst(compiler, STRBI | type | RT(reg) | RN(arg) | (argw << (10 - shift))); 846 } 847 848 if (argw <= 0xffffff) { 849 FAIL_IF(push_inst(compiler, ADDI | (1 << 22) | RD(tmp_reg) | RN(arg) | ((argw >> 12) << 10))); 850 851 argw = ((argw & 0xfff) >> shift); 852 return push_inst(compiler, STRBI | type | RT(reg) | RN(tmp_reg) | (argw << 10)); 853 } 854 } 855 856 if (argw <= 255 && argw >= -256) 857 return push_inst(compiler, STURBI | type | RT(reg) | RN(arg) | ((argw & 0x1ff) << 12)); 858 859 FAIL_IF(load_immediate(compiler, tmp_reg, argw)); 860 861 return push_inst(compiler, STRB | type | RT(reg) | RN(arg) | RM(tmp_reg)); 862 } 863 864 /* --------------------------------------------------------------------- */ 865 /* Entry, exit */ 866 /* --------------------------------------------------------------------- */ 867 868 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_enter(struct sljit_compiler *compiler, 869 sljit_s32 options, sljit_s32 arg_types, sljit_s32 scratches, sljit_s32 saveds, 870 sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size) 871 { 872 sljit_s32 args, i, tmp, offs, prev, saved_regs_size; 873 874 CHECK_ERROR(); 875 CHECK(check_sljit_emit_enter(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size)); 876 set_emit_enter(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size); 877 878 saved_regs_size = GET_SAVED_REGISTERS_SIZE(scratches, saveds, 2); 879 if (saved_regs_size & 0x8) 880 saved_regs_size += sizeof(sljit_sw); 881 882 local_size = (local_size + 15) & ~0xf; 883 compiler->local_size = local_size + saved_regs_size; 884 885 FAIL_IF(push_inst(compiler, STP_PRE | RT(TMP_FP) | RT2(TMP_LR) 886 | RN(SLJIT_SP) | ((-(saved_regs_size >> 3) & 0x7f) << 15))); 887 888 #ifdef _WIN32 889 if (local_size >= 4096) 890 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_REG1) | RN(SLJIT_SP) | (1 << 10) | (1 << 22))); 891 else if (local_size > 256) 892 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_REG1) | RN(SLJIT_SP) | (local_size << 10))); 893 #endif 894 895 tmp = saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - saveds) : SLJIT_FIRST_SAVED_REG; 896 prev = -1; 897 offs = 2 << 15; 898 for (i = SLJIT_S0; i >= tmp; i--) { 899 if (prev == -1) { 900 prev = i; 901 continue; 902 } 903 FAIL_IF(push_inst(compiler, STP | RT(prev) | RT2(i) | RN(SLJIT_SP) | offs)); 904 offs += 2 << 15; 905 prev = -1; 906 } 907 908 for (i = scratches; i >= SLJIT_FIRST_SAVED_REG; i--) { 909 if (prev == -1) { 910 prev = i; 911 continue; 912 } 913 FAIL_IF(push_inst(compiler, STP | RT(prev) | RT2(i) | RN(SLJIT_SP) | offs)); 914 offs += 2 << 15; 915 prev = -1; 916 } 917 918 if (prev != -1) 919 FAIL_IF(push_inst(compiler, STRI | RT(prev) | RN(SLJIT_SP) | (offs >> 5))); 920 921 922 FAIL_IF(push_inst(compiler, ADDI | RD(TMP_FP) | RN(SLJIT_SP) | (0 << 10))); 923 924 args = get_arg_count(arg_types); 925 926 if (args >= 1) 927 FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S0) | RN(TMP_ZERO) | RM(SLJIT_R0))); 928 if (args >= 2) 929 FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S1) | RN(TMP_ZERO) | RM(SLJIT_R1))); 930 if (args >= 3) 931 FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S2) | RN(TMP_ZERO) | RM(SLJIT_R2))); 932 933 #ifdef _WIN32 934 if (local_size >= 4096) { 935 if (local_size < 4 * 4096) { 936 /* No need for a loop. */ 937 if (local_size >= 2 * 4096) { 938 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 939 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_REG1) | RN(TMP_REG1) | (1 << 10) | (1 << 22))); 940 local_size -= 4096; 941 } 942 943 if (local_size >= 2 * 4096) { 944 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 945 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_REG1) | RN(TMP_REG1) | (1 << 10) | (1 << 22))); 946 local_size -= 4096; 947 } 948 949 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 950 local_size -= 4096; 951 } 952 else { 953 FAIL_IF(push_inst(compiler, MOVZ | RD(TMP_REG2) | (((local_size >> 12) - 1) << 5))); 954 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 955 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_REG1) | RN(TMP_REG1) | (1 << 10) | (1 << 22))); 956 FAIL_IF(push_inst(compiler, SUBI | (1 << 29) | RD(TMP_REG2) | RN(TMP_REG2) | (1 << 10))); 957 FAIL_IF(push_inst(compiler, B_CC | ((((sljit_ins) -3) & 0x7ffff) << 5) | 0x1 /* not-equal */)); 958 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 959 960 local_size &= 0xfff; 961 } 962 963 if (local_size > 256) { 964 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_REG1) | RN(TMP_REG1) | (local_size << 10))); 965 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 966 } 967 else if (local_size > 0) 968 FAIL_IF(push_inst(compiler, LDR_PRE | RT(TMP_ZERO) | RN(TMP_REG1) | ((-local_size & 0x1ff) << 12))); 969 970 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(TMP_REG1) | (0 << 10))); 971 } 972 else if (local_size > 256) { 973 FAIL_IF(push_inst(compiler, LDRI | RT(TMP_ZERO) | RN(TMP_REG1))); 974 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(TMP_REG1) | (0 << 10))); 975 } 976 else if (local_size > 0) 977 FAIL_IF(push_inst(compiler, LDR_PRE | RT(TMP_ZERO) | RN(SLJIT_SP) | ((-local_size & 0x1ff) << 12))); 978 979 #else /* !_WIN32 */ 980 981 /* The local_size does not include saved registers size. */ 982 if (local_size > 0xfff) { 983 FAIL_IF(push_inst(compiler, SUBI | RD(SLJIT_SP) | RN(SLJIT_SP) | ((local_size >> 12) << 10) | (1 << 22))); 984 local_size &= 0xfff; 985 } 986 if (local_size != 0) 987 FAIL_IF(push_inst(compiler, SUBI | RD(SLJIT_SP) | RN(SLJIT_SP) | (local_size << 10))); 988 989 #endif /* _WIN32 */ 990 991 return SLJIT_SUCCESS; 992 } 993 994 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_set_context(struct sljit_compiler *compiler, 995 sljit_s32 options, sljit_s32 arg_types, sljit_s32 scratches, sljit_s32 saveds, 996 sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size) 997 { 998 sljit_s32 saved_regs_size; 999 1000 CHECK_ERROR(); 1001 CHECK(check_sljit_set_context(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size)); 1002 set_set_context(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size); 1003 1004 saved_regs_size = GET_SAVED_REGISTERS_SIZE(scratches, saveds, 2); 1005 if (saved_regs_size & 0x8) 1006 saved_regs_size += sizeof(sljit_sw); 1007 1008 compiler->local_size = saved_regs_size + ((local_size + 15) & ~0xf); 1009 return SLJIT_SUCCESS; 1010 } 1011 1012 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_return(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src, sljit_sw srcw) 1013 { 1014 sljit_s32 local_size; 1015 sljit_s32 i, tmp, offs, prev, saved_regs_size; 1016 1017 CHECK_ERROR(); 1018 CHECK(check_sljit_emit_return(compiler, op, src, srcw)); 1019 1020 FAIL_IF(emit_mov_before_return(compiler, op, src, srcw)); 1021 1022 saved_regs_size = GET_SAVED_REGISTERS_SIZE(compiler->scratches, compiler->saveds, 2); 1023 if (saved_regs_size & 0x8) 1024 saved_regs_size += sizeof(sljit_sw); 1025 1026 local_size = compiler->local_size - saved_regs_size; 1027 1028 /* Load LR as early as possible. */ 1029 if (local_size == 0) 1030 FAIL_IF(push_inst(compiler, LDP | RT(TMP_FP) | RT2(TMP_LR) | RN(SLJIT_SP))); 1031 else if (local_size < 63 * sizeof(sljit_sw)) { 1032 FAIL_IF(push_inst(compiler, LDP_PRE | RT(TMP_FP) | RT2(TMP_LR) 1033 | RN(SLJIT_SP) | (local_size << (15 - 3)))); 1034 } 1035 else { 1036 if (local_size > 0xfff) { 1037 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(SLJIT_SP) | ((local_size >> 12) << 10) | (1 << 22))); 1038 local_size &= 0xfff; 1039 } 1040 if (local_size) 1041 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(SLJIT_SP) | (local_size << 10))); 1042 1043 FAIL_IF(push_inst(compiler, LDP | RT(TMP_FP) | RT2(TMP_LR) | RN(SLJIT_SP))); 1044 } 1045 1046 tmp = compiler->saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - compiler->saveds) : SLJIT_FIRST_SAVED_REG; 1047 prev = -1; 1048 offs = 2 << 15; 1049 for (i = SLJIT_S0; i >= tmp; i--) { 1050 if (prev == -1) { 1051 prev = i; 1052 continue; 1053 } 1054 FAIL_IF(push_inst(compiler, LDP | RT(prev) | RT2(i) | RN(SLJIT_SP) | offs)); 1055 offs += 2 << 15; 1056 prev = -1; 1057 } 1058 1059 for (i = compiler->scratches; i >= SLJIT_FIRST_SAVED_REG; i--) { 1060 if (prev == -1) { 1061 prev = i; 1062 continue; 1063 } 1064 FAIL_IF(push_inst(compiler, LDP | RT(prev) | RT2(i) | RN(SLJIT_SP) | offs)); 1065 offs += 2 << 15; 1066 prev = -1; 1067 } 1068 1069 if (prev != -1) 1070 FAIL_IF(push_inst(compiler, LDRI | RT(prev) | RN(SLJIT_SP) | (offs >> 5))); 1071 1072 /* These two can be executed in parallel. */ 1073 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(SLJIT_SP) | (saved_regs_size << 10))); 1074 return push_inst(compiler, RET | RN(TMP_LR)); 1075 } 1076 1077 /* --------------------------------------------------------------------- */ 1078 /* Operators */ 1079 /* --------------------------------------------------------------------- */ 1080 1081 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op0(struct sljit_compiler *compiler, sljit_s32 op) 1082 { 1083 sljit_ins inv_bits = (op & SLJIT_I32_OP) ? (1 << 31) : 0; 1084 1085 CHECK_ERROR(); 1086 CHECK(check_sljit_emit_op0(compiler, op)); 1087 1088 op = GET_OPCODE(op); 1089 switch (op) { 1090 case SLJIT_BREAKPOINT: 1091 return push_inst(compiler, BRK); 1092 case SLJIT_NOP: 1093 return push_inst(compiler, NOP); 1094 case SLJIT_LMUL_UW: 1095 case SLJIT_LMUL_SW: 1096 FAIL_IF(push_inst(compiler, ORR | RD(TMP_REG1) | RN(TMP_ZERO) | RM(SLJIT_R0))); 1097 FAIL_IF(push_inst(compiler, MADD | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1) | RT2(TMP_ZERO))); 1098 return push_inst(compiler, (op == SLJIT_LMUL_UW ? UMULH : SMULH) | RD(SLJIT_R1) | RN(TMP_REG1) | RM(SLJIT_R1)); 1099 case SLJIT_DIVMOD_UW: 1100 case SLJIT_DIVMOD_SW: 1101 FAIL_IF(push_inst(compiler, (ORR ^ inv_bits) | RD(TMP_REG1) | RN(TMP_ZERO) | RM(SLJIT_R0))); 1102 FAIL_IF(push_inst(compiler, ((op == SLJIT_DIVMOD_UW ? UDIV : SDIV) ^ inv_bits) | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1))); 1103 FAIL_IF(push_inst(compiler, (MADD ^ inv_bits) | RD(SLJIT_R1) | RN(SLJIT_R0) | RM(SLJIT_R1) | RT2(TMP_ZERO))); 1104 return push_inst(compiler, (SUB ^ inv_bits) | RD(SLJIT_R1) | RN(TMP_REG1) | RM(SLJIT_R1)); 1105 case SLJIT_DIV_UW: 1106 case SLJIT_DIV_SW: 1107 return push_inst(compiler, ((op == SLJIT_DIV_UW ? UDIV : SDIV) ^ inv_bits) | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1)); 1108 } 1109 1110 return SLJIT_SUCCESS; 1111 } 1112 1113 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op1(struct sljit_compiler *compiler, sljit_s32 op, 1114 sljit_s32 dst, sljit_sw dstw, 1115 sljit_s32 src, sljit_sw srcw) 1116 { 1117 sljit_s32 dst_r, flags, mem_flags; 1118 sljit_s32 op_flags = GET_ALL_FLAGS(op); 1119 1120 CHECK_ERROR(); 1121 CHECK(check_sljit_emit_op1(compiler, op, dst, dstw, src, srcw)); 1122 ADJUST_LOCAL_OFFSET(dst, dstw); 1123 ADJUST_LOCAL_OFFSET(src, srcw); 1124 1125 if (dst == SLJIT_UNUSED && !HAS_FLAGS(op)) { 1126 if (op <= SLJIT_MOV_P && (src & SLJIT_MEM)) { 1127 SLJIT_ASSERT(reg_map[1] == 0 && reg_map[3] == 2 && reg_map[5] == 4); 1128 1129 if (op >= SLJIT_MOV_U8 && op <= SLJIT_MOV_S8) 1130 dst = 5; 1131 else if (op >= SLJIT_MOV_U16 && op <= SLJIT_MOV_S16) 1132 dst = 3; 1133 else 1134 dst = 1; 1135 1136 /* Signed word sized load is the prefetch instruction. */ 1137 return emit_op_mem(compiler, WORD_SIZE | SIGNED, dst, src, srcw, TMP_REG1); 1138 } 1139 return SLJIT_SUCCESS; 1140 } 1141 1142 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1; 1143 1144 op = GET_OPCODE(op); 1145 if (op >= SLJIT_MOV && op <= SLJIT_MOV_P) { 1146 /* Both operands are registers. */ 1147 if (dst_r != TMP_REG1 && FAST_IS_REG(src)) 1148 return emit_op_imm(compiler, op | ((op_flags & SLJIT_I32_OP) ? INT_OP : 0), dst_r, TMP_REG1, src); 1149 1150 switch (op) { 1151 case SLJIT_MOV: 1152 case SLJIT_MOV_P: 1153 mem_flags = WORD_SIZE; 1154 break; 1155 case SLJIT_MOV_U8: 1156 mem_flags = BYTE_SIZE; 1157 if (src & SLJIT_IMM) 1158 srcw = (sljit_u8)srcw; 1159 break; 1160 case SLJIT_MOV_S8: 1161 mem_flags = BYTE_SIZE | SIGNED; 1162 if (src & SLJIT_IMM) 1163 srcw = (sljit_s8)srcw; 1164 break; 1165 case SLJIT_MOV_U16: 1166 mem_flags = HALF_SIZE; 1167 if (src & SLJIT_IMM) 1168 srcw = (sljit_u16)srcw; 1169 break; 1170 case SLJIT_MOV_S16: 1171 mem_flags = HALF_SIZE | SIGNED; 1172 if (src & SLJIT_IMM) 1173 srcw = (sljit_s16)srcw; 1174 break; 1175 case SLJIT_MOV_U32: 1176 mem_flags = INT_SIZE; 1177 if (src & SLJIT_IMM) 1178 srcw = (sljit_u32)srcw; 1179 break; 1180 case SLJIT_MOV_S32: 1181 mem_flags = INT_SIZE | SIGNED; 1182 if (src & SLJIT_IMM) 1183 srcw = (sljit_s32)srcw; 1184 break; 1185 default: 1186 SLJIT_UNREACHABLE(); 1187 mem_flags = 0; 1188 break; 1189 } 1190 1191 if (src & SLJIT_IMM) 1192 FAIL_IF(emit_op_imm(compiler, SLJIT_MOV | ARG2_IMM, dst_r, TMP_REG1, srcw)); 1193 else if (!(src & SLJIT_MEM)) 1194 dst_r = src; 1195 else 1196 FAIL_IF(emit_op_mem(compiler, mem_flags, dst_r, src, srcw, TMP_REG1)); 1197 1198 if (dst & SLJIT_MEM) 1199 return emit_op_mem(compiler, mem_flags | STORE, dst_r, dst, dstw, TMP_REG2); 1200 return SLJIT_SUCCESS; 1201 } 1202 1203 flags = HAS_FLAGS(op_flags) ? SET_FLAGS : 0; 1204 mem_flags = WORD_SIZE; 1205 1206 if (op_flags & SLJIT_I32_OP) { 1207 flags |= INT_OP; 1208 mem_flags = INT_SIZE; 1209 } 1210 1211 if (dst == SLJIT_UNUSED) 1212 flags |= UNUSED_RETURN; 1213 1214 if (src & SLJIT_MEM) { 1215 FAIL_IF(emit_op_mem(compiler, mem_flags, TMP_REG2, src, srcw, TMP_REG2)); 1216 src = TMP_REG2; 1217 } 1218 1219 emit_op_imm(compiler, flags | op, dst_r, TMP_REG1, src); 1220 1221 if (SLJIT_UNLIKELY(dst & SLJIT_MEM)) 1222 return emit_op_mem(compiler, mem_flags | STORE, dst_r, dst, dstw, TMP_REG2); 1223 return SLJIT_SUCCESS; 1224 } 1225 1226 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, 1227 sljit_s32 dst, sljit_sw dstw, 1228 sljit_s32 src1, sljit_sw src1w, 1229 sljit_s32 src2, sljit_sw src2w) 1230 { 1231 sljit_s32 dst_r, flags, mem_flags; 1232 1233 CHECK_ERROR(); 1234 CHECK(check_sljit_emit_op2(compiler, op, dst, dstw, src1, src1w, src2, src2w)); 1235 ADJUST_LOCAL_OFFSET(dst, dstw); 1236 ADJUST_LOCAL_OFFSET(src1, src1w); 1237 ADJUST_LOCAL_OFFSET(src2, src2w); 1238 1239 if (dst == SLJIT_UNUSED && !HAS_FLAGS(op)) 1240 return SLJIT_SUCCESS; 1241 1242 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1; 1243 flags = HAS_FLAGS(op) ? SET_FLAGS : 0; 1244 mem_flags = WORD_SIZE; 1245 1246 if (op & SLJIT_I32_OP) { 1247 flags |= INT_OP; 1248 mem_flags = INT_SIZE; 1249 } 1250 1251 if (dst == SLJIT_UNUSED) 1252 flags |= UNUSED_RETURN; 1253 1254 if (src1 & SLJIT_MEM) { 1255 FAIL_IF(emit_op_mem(compiler, mem_flags, TMP_REG1, src1, src1w, TMP_REG1)); 1256 src1 = TMP_REG1; 1257 } 1258 1259 if (src2 & SLJIT_MEM) { 1260 FAIL_IF(emit_op_mem(compiler, mem_flags, TMP_REG2, src2, src2w, TMP_REG2)); 1261 src2 = TMP_REG2; 1262 } 1263 1264 if (src1 & SLJIT_IMM) 1265 flags |= ARG1_IMM; 1266 else 1267 src1w = src1; 1268 1269 if (src2 & SLJIT_IMM) 1270 flags |= ARG2_IMM; 1271 else 1272 src2w = src2; 1273 1274 emit_op_imm(compiler, flags | GET_OPCODE(op), dst_r, src1w, src2w); 1275 1276 if (dst & SLJIT_MEM) 1277 return emit_op_mem(compiler, mem_flags | STORE, dst_r, dst, dstw, TMP_REG2); 1278 return SLJIT_SUCCESS; 1279 } 1280 1281 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_register_index(sljit_s32 reg) 1282 { 1283 CHECK_REG_INDEX(check_sljit_get_register_index(reg)); 1284 return reg_map[reg]; 1285 } 1286 1287 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_float_register_index(sljit_s32 reg) 1288 { 1289 CHECK_REG_INDEX(check_sljit_get_float_register_index(reg)); 1290 return freg_map[reg]; 1291 } 1292 1293 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_custom(struct sljit_compiler *compiler, 1294 void *instruction, sljit_s32 size) 1295 { 1296 CHECK_ERROR(); 1297 CHECK(check_sljit_emit_op_custom(compiler, instruction, size)); 1298 1299 return push_inst(compiler, *(sljit_ins*)instruction); 1300 } 1301 1302 /* --------------------------------------------------------------------- */ 1303 /* Floating point operators */ 1304 /* --------------------------------------------------------------------- */ 1305 1306 static sljit_s32 emit_fop_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw) 1307 { 1308 sljit_u32 shift = MEM_SIZE_SHIFT(flags); 1309 sljit_ins type = (shift << 30); 1310 1311 SLJIT_ASSERT(arg & SLJIT_MEM); 1312 1313 if (!(flags & STORE)) 1314 type |= 0x00400000; 1315 1316 if (arg & OFFS_REG_MASK) { 1317 argw &= 3; 1318 if (argw == 0 || argw == shift) 1319 return push_inst(compiler, STR_FR | type | VT(reg) 1320 | RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw ? (1 << 12) : 0)); 1321 1322 FAIL_IF(push_inst(compiler, ADD | RD(TMP_REG1) | RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw << 10))); 1323 return push_inst(compiler, STR_FI | type | VT(reg) | RN(TMP_REG1)); 1324 } 1325 1326 arg &= REG_MASK; 1327 1328 if (arg == SLJIT_UNUSED) { 1329 FAIL_IF(load_immediate(compiler, TMP_REG1, argw & ~(0xfff << shift))); 1330 1331 argw = (argw >> shift) & 0xfff; 1332 1333 return push_inst(compiler, STR_FI | type | VT(reg) | RN(TMP_REG1) | (argw << 10)); 1334 } 1335 1336 if (argw >= 0 && (argw & ((1 << shift) - 1)) == 0) { 1337 if ((argw >> shift) <= 0xfff) 1338 return push_inst(compiler, STR_FI | type | VT(reg) | RN(arg) | (argw << (10 - shift))); 1339 1340 if (argw <= 0xffffff) { 1341 FAIL_IF(push_inst(compiler, ADDI | (1 << 22) | RD(TMP_REG1) | RN(arg) | ((argw >> 12) << 10))); 1342 1343 argw = ((argw & 0xfff) >> shift); 1344 return push_inst(compiler, STR_FI | type | VT(reg) | RN(TMP_REG1) | (argw << 10)); 1345 } 1346 } 1347 1348 if (argw <= 255 && argw >= -256) 1349 return push_inst(compiler, STUR_FI | type | VT(reg) | RN(arg) | ((argw & 0x1ff) << 12)); 1350 1351 FAIL_IF(load_immediate(compiler, TMP_REG1, argw)); 1352 return push_inst(compiler, STR_FR | type | VT(reg) | RN(arg) | RM(TMP_REG1)); 1353 } 1354 1355 static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler *compiler, sljit_s32 op, 1356 sljit_s32 dst, sljit_sw dstw, 1357 sljit_s32 src, sljit_sw srcw) 1358 { 1359 sljit_s32 dst_r = FAST_IS_REG(dst) ? dst : TMP_REG1; 1360 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1361 1362 if (GET_OPCODE(op) == SLJIT_CONV_S32_FROM_F64) 1363 inv_bits |= (1 << 31); 1364 1365 if (src & SLJIT_MEM) { 1366 emit_fop_mem(compiler, (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE, TMP_FREG1, src, srcw); 1367 src = TMP_FREG1; 1368 } 1369 1370 FAIL_IF(push_inst(compiler, (FCVTZS ^ inv_bits) | RD(dst_r) | VN(src))); 1371 1372 if (dst & SLJIT_MEM) 1373 return emit_op_mem(compiler, ((GET_OPCODE(op) == SLJIT_CONV_S32_FROM_F64) ? INT_SIZE : WORD_SIZE) | STORE, TMP_REG1, dst, dstw, TMP_REG2); 1374 return SLJIT_SUCCESS; 1375 } 1376 1377 static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler *compiler, sljit_s32 op, 1378 sljit_s32 dst, sljit_sw dstw, 1379 sljit_s32 src, sljit_sw srcw) 1380 { 1381 sljit_s32 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1; 1382 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1383 1384 if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) 1385 inv_bits |= (1 << 31); 1386 1387 if (src & SLJIT_MEM) { 1388 emit_op_mem(compiler, ((GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) ? INT_SIZE : WORD_SIZE), TMP_REG1, src, srcw, TMP_REG1); 1389 src = TMP_REG1; 1390 } else if (src & SLJIT_IMM) { 1391 #if (defined SLJIT_CONFIG_X86_64 && SLJIT_CONFIG_X86_64) 1392 if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) 1393 srcw = (sljit_s32)srcw; 1394 #endif 1395 FAIL_IF(load_immediate(compiler, TMP_REG1, srcw)); 1396 src = TMP_REG1; 1397 } 1398 1399 FAIL_IF(push_inst(compiler, (SCVTF ^ inv_bits) | VD(dst_r) | RN(src))); 1400 1401 if (dst & SLJIT_MEM) 1402 return emit_fop_mem(compiler, ((op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE) | STORE, TMP_FREG1, dst, dstw); 1403 return SLJIT_SUCCESS; 1404 } 1405 1406 static SLJIT_INLINE sljit_s32 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op, 1407 sljit_s32 src1, sljit_sw src1w, 1408 sljit_s32 src2, sljit_sw src2w) 1409 { 1410 sljit_s32 mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE; 1411 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1412 1413 if (src1 & SLJIT_MEM) { 1414 emit_fop_mem(compiler, mem_flags, TMP_FREG1, src1, src1w); 1415 src1 = TMP_FREG1; 1416 } 1417 1418 if (src2 & SLJIT_MEM) { 1419 emit_fop_mem(compiler, mem_flags, TMP_FREG2, src2, src2w); 1420 src2 = TMP_FREG2; 1421 } 1422 1423 return push_inst(compiler, (FCMP ^ inv_bits) | VN(src1) | VM(src2)); 1424 } 1425 1426 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop1(struct sljit_compiler *compiler, sljit_s32 op, 1427 sljit_s32 dst, sljit_sw dstw, 1428 sljit_s32 src, sljit_sw srcw) 1429 { 1430 sljit_s32 dst_r, mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE; 1431 sljit_ins inv_bits; 1432 1433 CHECK_ERROR(); 1434 1435 SLJIT_COMPILE_ASSERT((INT_SIZE ^ 0x1) == WORD_SIZE, must_be_one_bit_difference); 1436 SELECT_FOP1_OPERATION_WITH_CHECKS(compiler, op, dst, dstw, src, srcw); 1437 1438 inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1439 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1; 1440 1441 if (src & SLJIT_MEM) { 1442 emit_fop_mem(compiler, (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_F32) ? (mem_flags ^ 0x1) : mem_flags, dst_r, src, srcw); 1443 src = dst_r; 1444 } 1445 1446 switch (GET_OPCODE(op)) { 1447 case SLJIT_MOV_F64: 1448 if (src != dst_r) { 1449 if (dst_r != TMP_FREG1) 1450 FAIL_IF(push_inst(compiler, (FMOV ^ inv_bits) | VD(dst_r) | VN(src))); 1451 else 1452 dst_r = src; 1453 } 1454 break; 1455 case SLJIT_NEG_F64: 1456 FAIL_IF(push_inst(compiler, (FNEG ^ inv_bits) | VD(dst_r) | VN(src))); 1457 break; 1458 case SLJIT_ABS_F64: 1459 FAIL_IF(push_inst(compiler, (FABS ^ inv_bits) | VD(dst_r) | VN(src))); 1460 break; 1461 case SLJIT_CONV_F64_FROM_F32: 1462 FAIL_IF(push_inst(compiler, FCVT | ((op & SLJIT_F32_OP) ? (1 << 22) : (1 << 15)) | VD(dst_r) | VN(src))); 1463 break; 1464 } 1465 1466 if (dst & SLJIT_MEM) 1467 return emit_fop_mem(compiler, mem_flags | STORE, dst_r, dst, dstw); 1468 return SLJIT_SUCCESS; 1469 } 1470 1471 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, 1472 sljit_s32 dst, sljit_sw dstw, 1473 sljit_s32 src1, sljit_sw src1w, 1474 sljit_s32 src2, sljit_sw src2w) 1475 { 1476 sljit_s32 dst_r, mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE; 1477 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1478 1479 CHECK_ERROR(); 1480 CHECK(check_sljit_emit_fop2(compiler, op, dst, dstw, src1, src1w, src2, src2w)); 1481 ADJUST_LOCAL_OFFSET(dst, dstw); 1482 ADJUST_LOCAL_OFFSET(src1, src1w); 1483 ADJUST_LOCAL_OFFSET(src2, src2w); 1484 1485 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1; 1486 if (src1 & SLJIT_MEM) { 1487 emit_fop_mem(compiler, mem_flags, TMP_FREG1, src1, src1w); 1488 src1 = TMP_FREG1; 1489 } 1490 if (src2 & SLJIT_MEM) { 1491 emit_fop_mem(compiler, mem_flags, TMP_FREG2, src2, src2w); 1492 src2 = TMP_FREG2; 1493 } 1494 1495 switch (GET_OPCODE(op)) { 1496 case SLJIT_ADD_F64: 1497 FAIL_IF(push_inst(compiler, (FADD ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1498 break; 1499 case SLJIT_SUB_F64: 1500 FAIL_IF(push_inst(compiler, (FSUB ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1501 break; 1502 case SLJIT_MUL_F64: 1503 FAIL_IF(push_inst(compiler, (FMUL ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1504 break; 1505 case SLJIT_DIV_F64: 1506 FAIL_IF(push_inst(compiler, (FDIV ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1507 break; 1508 } 1509 1510 if (!(dst & SLJIT_MEM)) 1511 return SLJIT_SUCCESS; 1512 return emit_fop_mem(compiler, mem_flags | STORE, TMP_FREG1, dst, dstw); 1513 } 1514 1515 /* --------------------------------------------------------------------- */ 1516 /* Other instructions */ 1517 /* --------------------------------------------------------------------- */ 1518 1519 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_enter(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw) 1520 { 1521 CHECK_ERROR(); 1522 CHECK(check_sljit_emit_fast_enter(compiler, dst, dstw)); 1523 ADJUST_LOCAL_OFFSET(dst, dstw); 1524 1525 if (FAST_IS_REG(dst)) 1526 return push_inst(compiler, ORR | RD(dst) | RN(TMP_ZERO) | RM(TMP_LR)); 1527 1528 /* Memory. */ 1529 return emit_op_mem(compiler, WORD_SIZE | STORE, TMP_LR, dst, dstw, TMP_REG1); 1530 } 1531 1532 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_return(struct sljit_compiler *compiler, sljit_s32 src, sljit_sw srcw) 1533 { 1534 CHECK_ERROR(); 1535 CHECK(check_sljit_emit_fast_return(compiler, src, srcw)); 1536 ADJUST_LOCAL_OFFSET(src, srcw); 1537 1538 if (FAST_IS_REG(src)) 1539 FAIL_IF(push_inst(compiler, ORR | RD(TMP_LR) | RN(TMP_ZERO) | RM(src))); 1540 else 1541 FAIL_IF(emit_op_mem(compiler, WORD_SIZE, TMP_LR, src, srcw, TMP_REG1)); 1542 1543 return push_inst(compiler, RET | RN(TMP_LR)); 1544 } 1545 1546 /* --------------------------------------------------------------------- */ 1547 /* Conditional instructions */ 1548 /* --------------------------------------------------------------------- */ 1549 1550 static sljit_uw get_cc(sljit_s32 type) 1551 { 1552 switch (type) { 1553 case SLJIT_EQUAL: 1554 case SLJIT_MUL_NOT_OVERFLOW: 1555 case SLJIT_EQUAL_F64: 1556 return 0x1; 1557 1558 case SLJIT_NOT_EQUAL: 1559 case SLJIT_MUL_OVERFLOW: 1560 case SLJIT_NOT_EQUAL_F64: 1561 return 0x0; 1562 1563 case SLJIT_LESS: 1564 case SLJIT_LESS_F64: 1565 return 0x2; 1566 1567 case SLJIT_GREATER_EQUAL: 1568 case SLJIT_GREATER_EQUAL_F64: 1569 return 0x3; 1570 1571 case SLJIT_GREATER: 1572 case SLJIT_GREATER_F64: 1573 return 0x9; 1574 1575 case SLJIT_LESS_EQUAL: 1576 case SLJIT_LESS_EQUAL_F64: 1577 return 0x8; 1578 1579 case SLJIT_SIG_LESS: 1580 return 0xa; 1581 1582 case SLJIT_SIG_GREATER_EQUAL: 1583 return 0xb; 1584 1585 case SLJIT_SIG_GREATER: 1586 return 0xd; 1587 1588 case SLJIT_SIG_LESS_EQUAL: 1589 return 0xc; 1590 1591 case SLJIT_OVERFLOW: 1592 case SLJIT_UNORDERED_F64: 1593 return 0x7; 1594 1595 case SLJIT_NOT_OVERFLOW: 1596 case SLJIT_ORDERED_F64: 1597 return 0x6; 1598 1599 default: 1600 SLJIT_UNREACHABLE(); 1601 return 0xe; 1602 } 1603 } 1604 1605 SLJIT_API_FUNC_ATTRIBUTE struct sljit_label* sljit_emit_label(struct sljit_compiler *compiler) 1606 { 1607 struct sljit_label *label; 1608 1609 CHECK_ERROR_PTR(); 1610 CHECK_PTR(check_sljit_emit_label(compiler)); 1611 1612 if (compiler->last_label && compiler->last_label->size == compiler->size) 1613 return compiler->last_label; 1614 1615 label = (struct sljit_label*)ensure_abuf(compiler, sizeof(struct sljit_label)); 1616 PTR_FAIL_IF(!label); 1617 set_label(label, compiler); 1618 return label; 1619 } 1620 1621 SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_jump(struct sljit_compiler *compiler, sljit_s32 type) 1622 { 1623 struct sljit_jump *jump; 1624 1625 CHECK_ERROR_PTR(); 1626 CHECK_PTR(check_sljit_emit_jump(compiler, type)); 1627 1628 jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump)); 1629 PTR_FAIL_IF(!jump); 1630 set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP); 1631 type &= 0xff; 1632 1633 if (type < SLJIT_JUMP) { 1634 jump->flags |= IS_COND; 1635 PTR_FAIL_IF(push_inst(compiler, B_CC | (6 << 5) | get_cc(type))); 1636 } 1637 else if (type >= SLJIT_FAST_CALL) 1638 jump->flags |= IS_BL; 1639 1640 PTR_FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0)); 1641 jump->addr = compiler->size; 1642 PTR_FAIL_IF(push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(TMP_REG1))); 1643 1644 return jump; 1645 } 1646 1647 SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_call(struct sljit_compiler *compiler, sljit_s32 type, 1648 sljit_s32 arg_types) 1649 { 1650 CHECK_ERROR_PTR(); 1651 CHECK_PTR(check_sljit_emit_call(compiler, type, arg_types)); 1652 1653 #if (defined SLJIT_VERBOSE && SLJIT_VERBOSE) \ 1654 || (defined SLJIT_ARGUMENT_CHECKS && SLJIT_ARGUMENT_CHECKS) 1655 compiler->skip_checks = 1; 1656 #endif 1657 1658 return sljit_emit_jump(compiler, type); 1659 } 1660 1661 static SLJIT_INLINE struct sljit_jump* emit_cmp_to0(struct sljit_compiler *compiler, sljit_s32 type, 1662 sljit_s32 src, sljit_sw srcw) 1663 { 1664 struct sljit_jump *jump; 1665 sljit_ins inv_bits = (type & SLJIT_I32_OP) ? (1 << 31) : 0; 1666 1667 SLJIT_ASSERT((type & 0xff) == SLJIT_EQUAL || (type & 0xff) == SLJIT_NOT_EQUAL); 1668 ADJUST_LOCAL_OFFSET(src, srcw); 1669 1670 jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump)); 1671 PTR_FAIL_IF(!jump); 1672 set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP); 1673 jump->flags |= IS_CBZ | IS_COND; 1674 1675 if (src & SLJIT_MEM) { 1676 PTR_FAIL_IF(emit_op_mem(compiler, inv_bits ? INT_SIZE : WORD_SIZE, TMP_REG1, src, srcw, TMP_REG1)); 1677 src = TMP_REG1; 1678 } 1679 else if (src & SLJIT_IMM) { 1680 PTR_FAIL_IF(load_immediate(compiler, TMP_REG1, srcw)); 1681 src = TMP_REG1; 1682 } 1683 1684 SLJIT_ASSERT(FAST_IS_REG(src)); 1685 1686 if ((type & 0xff) == SLJIT_EQUAL) 1687 inv_bits |= 1 << 24; 1688 1689 PTR_FAIL_IF(push_inst(compiler, (CBZ ^ inv_bits) | (6 << 5) | RT(src))); 1690 PTR_FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0)); 1691 jump->addr = compiler->size; 1692 PTR_FAIL_IF(push_inst(compiler, BR | RN(TMP_REG1))); 1693 return jump; 1694 } 1695 1696 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_ijump(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src, sljit_sw srcw) 1697 { 1698 struct sljit_jump *jump; 1699 1700 CHECK_ERROR(); 1701 CHECK(check_sljit_emit_ijump(compiler, type, src, srcw)); 1702 ADJUST_LOCAL_OFFSET(src, srcw); 1703 1704 if (!(src & SLJIT_IMM)) { 1705 if (src & SLJIT_MEM) { 1706 FAIL_IF(emit_op_mem(compiler, WORD_SIZE, TMP_REG1, src, srcw, TMP_REG1)); 1707 src = TMP_REG1; 1708 } 1709 return push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(src)); 1710 } 1711 1712 /* These jumps are converted to jump/call instructions when possible. */ 1713 jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump)); 1714 FAIL_IF(!jump); 1715 set_jump(jump, compiler, JUMP_ADDR | ((type >= SLJIT_FAST_CALL) ? IS_BL : 0)); 1716 jump->u.target = srcw; 1717 1718 FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0)); 1719 jump->addr = compiler->size; 1720 return push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(TMP_REG1)); 1721 } 1722 1723 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_icall(struct sljit_compiler *compiler, sljit_s32 type, 1724 sljit_s32 arg_types, 1725 sljit_s32 src, sljit_sw srcw) 1726 { 1727 CHECK_ERROR(); 1728 CHECK(check_sljit_emit_icall(compiler, type, arg_types, src, srcw)); 1729 1730 #if (defined SLJIT_VERBOSE && SLJIT_VERBOSE) \ 1731 || (defined SLJIT_ARGUMENT_CHECKS && SLJIT_ARGUMENT_CHECKS) 1732 compiler->skip_checks = 1; 1733 #endif 1734 1735 return sljit_emit_ijump(compiler, type, src, srcw); 1736 } 1737 1738 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_flags(struct sljit_compiler *compiler, sljit_s32 op, 1739 sljit_s32 dst, sljit_sw dstw, 1740 sljit_s32 type) 1741 { 1742 sljit_s32 dst_r, src_r, flags, mem_flags; 1743 sljit_ins cc; 1744 1745 CHECK_ERROR(); 1746 CHECK(check_sljit_emit_op_flags(compiler, op, dst, dstw, type)); 1747 ADJUST_LOCAL_OFFSET(dst, dstw); 1748 1749 cc = get_cc(type & 0xff); 1750 dst_r = FAST_IS_REG(dst) ? dst : TMP_REG1; 1751 1752 if (GET_OPCODE(op) < SLJIT_ADD) { 1753 FAIL_IF(push_inst(compiler, CSINC | (cc << 12) | RD(dst_r) | RN(TMP_ZERO) | RM(TMP_ZERO))); 1754 1755 if (dst_r == TMP_REG1) { 1756 mem_flags = (GET_OPCODE(op) == SLJIT_MOV ? WORD_SIZE : INT_SIZE) | STORE; 1757 return emit_op_mem(compiler, mem_flags, TMP_REG1, dst, dstw, TMP_REG2); 1758 } 1759 1760 return SLJIT_SUCCESS; 1761 } 1762 1763 flags = HAS_FLAGS(op) ? SET_FLAGS : 0; 1764 mem_flags = WORD_SIZE; 1765 1766 if (op & SLJIT_I32_OP) { 1767 flags |= INT_OP; 1768 mem_flags = INT_SIZE; 1769 } 1770 1771 src_r = dst; 1772 1773 if (dst & SLJIT_MEM) { 1774 FAIL_IF(emit_op_mem(compiler, mem_flags, TMP_REG1, dst, dstw, TMP_REG1)); 1775 src_r = TMP_REG1; 1776 } 1777 1778 FAIL_IF(push_inst(compiler, CSINC | (cc << 12) | RD(TMP_REG2) | RN(TMP_ZERO) | RM(TMP_ZERO))); 1779 emit_op_imm(compiler, flags | GET_OPCODE(op), dst_r, src_r, TMP_REG2); 1780 1781 if (dst & SLJIT_MEM) 1782 return emit_op_mem(compiler, mem_flags | STORE, TMP_REG1, dst, dstw, TMP_REG2); 1783 return SLJIT_SUCCESS; 1784 } 1785 1786 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_cmov(struct sljit_compiler *compiler, sljit_s32 type, 1787 sljit_s32 dst_reg, 1788 sljit_s32 src, sljit_sw srcw) 1789 { 1790 sljit_ins inv_bits = (dst_reg & SLJIT_I32_OP) ? (1 << 31) : 0; 1791 sljit_ins cc; 1792 1793 CHECK_ERROR(); 1794 CHECK(check_sljit_emit_cmov(compiler, type, dst_reg, src, srcw)); 1795 1796 if (SLJIT_UNLIKELY(src & SLJIT_IMM)) { 1797 if (dst_reg & SLJIT_I32_OP) 1798 srcw = (sljit_s32)srcw; 1799 FAIL_IF(load_immediate(compiler, TMP_REG1, srcw)); 1800 src = TMP_REG1; 1801 srcw = 0; 1802 } 1803 1804 cc = get_cc(type & 0xff); 1805 dst_reg &= ~SLJIT_I32_OP; 1806 1807 return push_inst(compiler, (CSEL ^ inv_bits) | (cc << 12) | RD(dst_reg) | RN(dst_reg) | RM(src)); 1808 } 1809 1810 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_mem(struct sljit_compiler *compiler, sljit_s32 type, 1811 sljit_s32 reg, 1812 sljit_s32 mem, sljit_sw memw) 1813 { 1814 sljit_u32 sign = 0, inst; 1815 1816 CHECK_ERROR(); 1817 CHECK(check_sljit_emit_mem(compiler, type, reg, mem, memw)); 1818 1819 if ((mem & OFFS_REG_MASK) || (memw > 255 && memw < -256)) 1820 return SLJIT_ERR_UNSUPPORTED; 1821 1822 if (type & SLJIT_MEM_SUPP) 1823 return SLJIT_SUCCESS; 1824 1825 switch (type & 0xff) { 1826 case SLJIT_MOV: 1827 case SLJIT_MOV_P: 1828 inst = STURBI | (MEM_SIZE_SHIFT(WORD_SIZE) << 30) | 0x400; 1829 break; 1830 case SLJIT_MOV_S8: 1831 sign = 1; 1832 case SLJIT_MOV_U8: 1833 inst = STURBI | (MEM_SIZE_SHIFT(BYTE_SIZE) << 30) | 0x400; 1834 break; 1835 case SLJIT_MOV_S16: 1836 sign = 1; 1837 case SLJIT_MOV_U16: 1838 inst = STURBI | (MEM_SIZE_SHIFT(HALF_SIZE) << 30) | 0x400; 1839 break; 1840 case SLJIT_MOV_S32: 1841 sign = 1; 1842 case SLJIT_MOV_U32: 1843 inst = STURBI | (MEM_SIZE_SHIFT(INT_SIZE) << 30) | 0x400; 1844 break; 1845 default: 1846 SLJIT_UNREACHABLE(); 1847 inst = STURBI | (MEM_SIZE_SHIFT(WORD_SIZE) << 30) | 0x400; 1848 break; 1849 } 1850 1851 if (!(type & SLJIT_MEM_STORE)) 1852 inst |= sign ? 0x00800000 : 0x00400000; 1853 1854 if (type & SLJIT_MEM_PRE) 1855 inst |= 0x800; 1856 1857 return push_inst(compiler, inst | RT(reg) | RN(mem & REG_MASK) | ((memw & 0x1ff) << 12)); 1858 } 1859 1860 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fmem(struct sljit_compiler *compiler, sljit_s32 type, 1861 sljit_s32 freg, 1862 sljit_s32 mem, sljit_sw memw) 1863 { 1864 sljit_u32 inst; 1865 1866 CHECK_ERROR(); 1867 CHECK(check_sljit_emit_fmem(compiler, type, freg, mem, memw)); 1868 1869 if ((mem & OFFS_REG_MASK) || (memw > 255 && memw < -256)) 1870 return SLJIT_ERR_UNSUPPORTED; 1871 1872 if (type & SLJIT_MEM_SUPP) 1873 return SLJIT_SUCCESS; 1874 1875 inst = STUR_FI | 0x80000400; 1876 1877 if (!(type & SLJIT_F32_OP)) 1878 inst |= 0x40000000; 1879 1880 if (!(type & SLJIT_MEM_STORE)) 1881 inst |= 0x00400000; 1882 1883 if (type & SLJIT_MEM_PRE) 1884 inst |= 0x800; 1885 1886 return push_inst(compiler, inst | VT(freg) | RN(mem & REG_MASK) | ((memw & 0x1ff) << 12)); 1887 } 1888 1889 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_local_base(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw offset) 1890 { 1891 sljit_s32 dst_reg; 1892 sljit_ins ins; 1893 1894 CHECK_ERROR(); 1895 CHECK(check_sljit_get_local_base(compiler, dst, dstw, offset)); 1896 1897 SLJIT_ASSERT (SLJIT_LOCALS_OFFSET_BASE == 0); 1898 1899 dst_reg = FAST_IS_REG(dst) ? dst : TMP_REG1; 1900 1901 if (offset <= 0xffffff && offset >= -0xffffff) { 1902 ins = ADDI; 1903 if (offset < 0) { 1904 offset = -offset; 1905 ins = SUBI; 1906 } 1907 1908 if (offset <= 0xfff) 1909 FAIL_IF(push_inst(compiler, ins | RD(dst_reg) | RN(SLJIT_SP) | (offset << 10))); 1910 else { 1911 FAIL_IF(push_inst(compiler, ins | RD(dst_reg) | RN(SLJIT_SP) | ((offset & 0xfff000) >> (12 - 10)) | (1 << 22))); 1912 1913 offset &= 0xfff; 1914 if (offset != 0) 1915 FAIL_IF(push_inst(compiler, ins | RD(dst_reg) | RN(dst_reg) | (offset << 10))); 1916 } 1917 } 1918 else { 1919 FAIL_IF(load_immediate (compiler, dst_reg, offset)); 1920 /* Add extended register form. */ 1921 FAIL_IF(push_inst(compiler, ADDE | (0x3 << 13) | RD(dst_reg) | RN(SLJIT_SP) | RM(dst_reg))); 1922 } 1923 1924 if (SLJIT_UNLIKELY(dst & SLJIT_MEM)) 1925 return emit_op_mem(compiler, WORD_SIZE | STORE, dst_reg, dst, dstw, TMP_REG1); 1926 return SLJIT_SUCCESS; 1927 } 1928 1929 SLJIT_API_FUNC_ATTRIBUTE struct sljit_const* sljit_emit_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw init_value) 1930 { 1931 struct sljit_const *const_; 1932 sljit_s32 dst_r; 1933 1934 CHECK_ERROR_PTR(); 1935 CHECK_PTR(check_sljit_emit_const(compiler, dst, dstw, init_value)); 1936 ADJUST_LOCAL_OFFSET(dst, dstw); 1937 1938 const_ = (struct sljit_const*)ensure_abuf(compiler, sizeof(struct sljit_const)); 1939 PTR_FAIL_IF(!const_); 1940 set_const(const_, compiler); 1941 1942 dst_r = FAST_IS_REG(dst) ? dst : TMP_REG1; 1943 PTR_FAIL_IF(emit_imm64_const(compiler, dst_r, init_value)); 1944 1945 if (dst & SLJIT_MEM) 1946 PTR_FAIL_IF(emit_op_mem(compiler, WORD_SIZE | STORE, dst_r, dst, dstw, TMP_REG2)); 1947 return const_; 1948 } 1949 1950 SLJIT_API_FUNC_ATTRIBUTE void sljit_set_jump_addr(sljit_uw addr, sljit_uw new_target, sljit_sw executable_offset) 1951 { 1952 sljit_ins* inst = (sljit_ins*)addr; 1953 modify_imm64_const(inst, new_target); 1954 inst = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(inst, executable_offset); 1955 SLJIT_CACHE_FLUSH(inst, inst + 4); 1956 } 1957 1958 SLJIT_API_FUNC_ATTRIBUTE void sljit_set_const(sljit_uw addr, sljit_sw new_constant, sljit_sw executable_offset) 1959 { 1960 sljit_ins* inst = (sljit_ins*)addr; 1961 modify_imm64_const(inst, new_constant); 1962 inst = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(inst, executable_offset); 1963 SLJIT_CACHE_FLUSH(inst, inst + 4); 1964 } 1965