/external/chromium_org/v8/test/cctest/ |
test-random.cc | 60 double crankshaft_value = HeapNumber::cast(*value)->value(); 64 double runtime_value = HeapNumber::cast(*value)->value(); 106 double crankshaft_value = HeapNumber::cast(*value)->value();
|
test-code-stubs.cc | 50 HeapNumber::kExponentBias);
|
test-code-stubs-x64.cc | 57 source_reg.is(rsp) ? 0 : (HeapNumber::kValueOffset - kSmiTagSize);
|
/external/v8/test/cctest/ |
test-random.cc | 63 double crankshaft_value = HeapNumber::cast(*value)->value(); 67 double runtime_value = HeapNumber::cast(*value)->value(); 107 double crankshaft_value = HeapNumber::cast(*value)->value();
|
/external/chromium_org/v8/src/ |
v8conversions.h | 66 double value = HeapNumber::cast(number)->value();
|
/external/v8/src/mips/ |
code-stubs-mips.cc | 472 STATIC_ASSERT(HeapNumber::kSignMask == 0x80000000u); 473 __ And(exponent, source_, Operand(HeapNumber::kSignMask)); 485 HeapNumber::kExponentBias << HeapNumber::kExponentShift; 499 __ li(mantissa, Operand(31 + HeapNumber::kExponentBias)); 501 __ sll(mantissa, mantissa, HeapNumber::kExponentShift); 509 __ sll(mantissa, source_, HeapNumber::kMantissaBitsInTopWord); 511 __ srl(source_, source_, 32 - HeapNumber::kMantissaBitsInTopWord); 595 // Load the double from tagged HeapNumber to double register. 597 // ARM uses a workaround here because of the unaligned HeapNumber 3589 const Register heapnumber = v0; local [all...] |
stub-cache-mips.cc | [all...] |
/external/v8/src/arm/ |
stub-cache-arm.cc | [all...] |
code-stubs-arm.cc | 471 STATIC_ASSERT(HeapNumber::kSignMask == 0x80000000u); 472 __ and_(exponent, source_, Operand(HeapNumber::kSignMask), SetCC); 484 HeapNumber::kExponentBias << HeapNumber::kExponentShift; 496 // divide the constant 31 + HeapNumber::kExponentBias, 0x41d, into two parts 499 __ rsb(mantissa, zeros_, Operand(31 + HeapNumber::kExponentBias - fudge)); 503 Operand(mantissa, LSL, HeapNumber::kExponentShift)); 509 __ mov(mantissa, Operand(source_, LSL, HeapNumber::kMantissaBitsInTopWord)); 513 Operand(source_, LSR, 32 - HeapNumber::kMantissaBitsInTopWord)); 595 // Load the double from tagged HeapNumber to double register 3452 const Register heapnumber = r0; local [all...] |
macro-assembler-arm.cc | [all...] |
/external/v8/src/ia32/ |
code-stubs-ia32.cc | 500 __ fld_d(FieldOperand(argument, HeapNumber::kValueOffset)); 648 // Test if operands are numbers (smi or HeapNumber objects), and load 679 __ mov(scratch, FieldOperand(source, HeapNumber::kExponentOffset)); 682 __ and_(scratch2, HeapNumber::kExponentMask); 687 (HeapNumber::kExponentBias + 63) << HeapNumber::kExponentShift; 691 __ fld_d(FieldOperand(source, HeapNumber::kValueOffset)); 707 (HeapNumber::kExponentBias + 30) << HeapNumber::kExponentShift; 720 (HeapNumber::kExponentBias + 31) << HeapNumber::kExponentShift [all...] |
stub-cache-ia32.cc | [all...] |
/external/v8/src/ |
v8.cc | 238 HeapNumber::cast(heap_number)->set_value(r.double_value);
|
/external/chromium_org/v8/test/mjsunit/regress/ |
regress-137.js | 34 var base = base - (rem / 16); // base is now HeapNumber with valid Smi value.
|
regress-155924.js | 28 // A HeapNumber with certain bits in the mantissa of the floating point
|
/external/v8/test/mjsunit/regress/ |
regress-137.js | 34 var base = base - (rem / 16); // base is now HeapNumber with valid Smi value.
|
regress-155924.js | 28 // A HeapNumber with certain bits in the mantissa of the floating point
|
/external/v8/src/x64/ |
code-stubs-x64.cc | 482 __ ucomisd(xmm0, FieldOperand(argument, HeapNumber::kValueOffset)); 578 // This fails if either argument is not a Smi nor a HeapNumber, 579 // or if it's a HeapNumber with a value that can't be converted 612 __ movq(double_value, FieldOperand(source, HeapNumber::kValueOffset)); 619 __ shr(double_exponent, Immediate(64 - HeapNumber::kExponentBits)); 620 __ subl(double_exponent, Immediate(HeapNumber::kExponentBias)); 652 __ leal(rcx, Operand(double_exponent, -HeapNumber::kMantissaBits - 1)); 658 __ leal(rcx, Operand(double_exponent, -HeapNumber::kMantissaBits - 1)); 804 __ xor_(FieldOperand(rax, HeapNumber::kValueOffset), kScratchRegister); 824 __ movq(rdx, FieldOperand(rax, HeapNumber::kValueOffset)) [all...] |
/external/chromium_org/v8/src/mips/ |
code-stubs-mips.cc | 595 STATIC_ASSERT(HeapNumber::kSignMask == 0x80000000u); 596 __ And(exponent, source_, Operand(HeapNumber::kSignMask)); 608 HeapNumber::kExponentBias << HeapNumber::kExponentShift; 622 __ li(mantissa, Operand(31 + HeapNumber::kExponentBias)); 624 __ sll(mantissa, mantissa, HeapNumber::kExponentShift); 632 __ sll(mantissa, source_, HeapNumber::kMantissaBitsInTopWord); 634 __ srl(source_, source_, 32 - HeapNumber::kMantissaBitsInTopWord); 681 // Load the double from tagged HeapNumber to double register. 683 // ARM uses a workaround here because of the unaligned HeapNumber 2833 const Register heapnumber = v0; local [all...] |
stub-cache-mips.cc | 495 __ ldc1(f4, FieldMemOperand(value_reg, HeapNumber::kValueOffset)); 498 __ sdc1(f4, FieldMemOperand(storage_reg, HeapNumber::kValueOffset)); 667 __ ldc1(f4, FieldMemOperand(value_reg, HeapNumber::kValueOffset)); 670 __ sdc1(f4, FieldMemOperand(scratch1, HeapNumber::kValueOffset)); [all...] |
/external/chromium_org/v8/src/ia32/ |
stub-cache-ia32.cc | [all...] |
code-stubs-ia32.cc | 621 // Test if operands are numbers (smi or HeapNumber objects), and load 686 __ and_(ecx, HeapNumber::kExponentMask); 687 __ shr(ecx, HeapNumber::kExponentShift); 688 __ lea(result_reg, MemOperand(ecx, -HeapNumber::kExponentBias)); 689 __ cmp(result_reg, Immediate(HeapNumber::kMantissaBits)); 693 int delta = HeapNumber::kExponentBias + Double::kPhysicalSignificandSize; 788 __ movdbl(xmm0, FieldOperand(source, HeapNumber::kValueOffset)); [all...] |
lithium-codegen-ia32.cc | [all...] |
/external/chromium_org/v8/src/arm/ |
code-stubs-arm.cc | 592 STATIC_ASSERT(HeapNumber::kSignMask == 0x80000000u); 593 __ and_(exponent, source_, Operand(HeapNumber::kSignMask), SetCC); 605 HeapNumber::kExponentBias << HeapNumber::kExponentShift; 615 // divide the constant 31 + HeapNumber::kExponentBias, 0x41d, into two parts 618 __ rsb(mantissa, zeros_, Operand(31 + HeapNumber::kExponentBias - fudge)); 622 Operand(mantissa, LSL, HeapNumber::kExponentShift)); 628 __ mov(mantissa, Operand(source_, LSL, HeapNumber::kMantissaBitsInTopWord)); 632 Operand(source_, LSR, 32 - HeapNumber::kMantissaBitsInTopWord)); 667 STATIC_ASSERT(HeapNumber::kSignMask == 0x80000000u) 2525 const Register heapnumber = r0; local [all...] |
stub-cache-arm.cc | 505 __ vldr(d0, FieldMemOperand(value_reg, HeapNumber::kValueOffset)); 508 __ vstr(d0, FieldMemOperand(storage_reg, HeapNumber::kValueOffset)); 676 __ vldr(d0, FieldMemOperand(value_reg, HeapNumber::kValueOffset)); 679 __ vstr(d0, FieldMemOperand(scratch1, HeapNumber::kValueOffset)); [all...] |