/external/llvm/lib/Target/X86/ |
README.txt | 497 movsd 24(%esp), %xmm1 789 movsd 24(%esp), %xmm0 790 movsd %xmm0, 8(%esp) 885 movsd (%esp), %xmm0 887 movsd %xmm0, (%esp) 1016 movsd 176(%esp), %xmm2 1034 movsd 152(%esp), %xmm1 1036 movsd %xmm1, 152(%esp) 1041 movsd 152(%esp), %xmm0 1044 movsd %xmm0, 152(%esp [all...] |
README-SSE.txt | 390 movsd 16(%esp), %xmm0 391 movsd %xmm0, (%esp) 716 Consider using movlps instead of movsd to implement (scalar_to_vector (loadf64)) 717 when code size is critical. movlps is slower than movsd on core2 but it's one 749 movsd %xmm0, (%esp)
|
X86TargetTransformInfo.cpp | 463 // These are lowered into movsd. 486 {ISD::VECTOR_SHUFFLE, MVT::v2i64, 1}, // movsd 487 {ISD::VECTOR_SHUFFLE, MVT::v2f64, 1}, // movsd 504 {ISD::VECTOR_SHUFFLE, MVT::v2i64, 1}, // movsd 505 {ISD::VECTOR_SHUFFLE, MVT::v2f64, 1}, // movsd [all...] |
/art/compiler/utils/x86_64/ |
assembler_x86_64.h | 415 void movsd(XmmRegister dst, const Address& src); 416 void movsd(const Address& dst, XmmRegister src); 417 void movsd(XmmRegister dst, XmmRegister src); [all...] |
assembler_x86_64_test.cc | [all...] |
/prebuilts/go/darwin-x86/src/cmd/internal/rsc.io/x86/x86asm/ |
gnu.go | 396 case MOVSB, MOVSW, MOVSD, MOVSQ, OUTSB, OUTSW, OUTSD: 821 MOVSD: "movsl", 822 MOVSD_XMM: "movsd",
|
/prebuilts/go/linux-x86/src/cmd/internal/rsc.io/x86/x86asm/ |
gnu.go | 396 case MOVSB, MOVSW, MOVSD, MOVSQ, OUTSB, OUTSW, OUTSD: 821 MOVSD: "movsl", 822 MOVSD_XMM: "movsd",
|
/external/llvm/test/CodeGen/X86/ |
fp-logic.ll | 238 ; CHECK-NEXT: movsd {{.*#+}} xmm1 = mem[0],zero
|
sink-hoist.ll | 43 ; CHECK: movsd
|
vector-shuffle-combining.ll | [all...] |
vector-shift-shl-128.ll | 23 ; SSE2-NEXT: movsd {{.*#+}} xmm2 = xmm0[0],xmm2[1] 66 ; X32-SSE-NEXT: movsd {{.*#+}} xmm2 = xmm0[0],xmm2[1] 649 ; SSE2-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1] 688 ; X32-SSE-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1] [all...] |
/external/v8/src/ia32/ |
macro-assembler-ia32.cc | 149 movsd(Operand(esp, (i - 1) * kDoubleSize), reg); 160 movsd(reg, Operand(esp, (i - 1) * kDoubleSize)); 279 movsd(MemOperand(esp, 0), input_reg); 353 movsd(xmm0, FieldOperand(input_reg, HeapNumber::kValueOffset)); 370 movsd(MemOperand(esp, 0), xmm0); 883 movsd(FieldOperand(elements, key, times_4, [all...] |
/art/compiler/optimizing/ |
code_generator_x86.cc | 752 __ movsd(Address(ESP, stack_index), XmmRegister(reg_id)); 757 __ movsd(XmmRegister(reg_id), Address(ESP, stack_index)); [all...] |
code_generator_x86_64.cc | [all...] |
/art/runtime/interpreter/mterp/out/ |
mterp_x86_64.S | [all...] |
mterp_x86.S | [all...] |
/external/v8/src/crankshaft/x64/ |
lithium-codegen-x64.cc | 91 __ Movsd(MemOperand(rsp, count * kDoubleSize), 107 __ Movsd(XMMRegister::from_code(save_iterator.Current()), [all...] |
/external/v8/test/cctest/ |
test-disasm-ia32.cc | 437 __ movsd(xmm1, Operand(ebx, ecx, times_4, 10000)); 438 __ movsd(Operand(ebx, ecx, times_4, 10000), xmm1);
|
test-disasm-x64.cc | 431 __ movsd(xmm1, Operand(rbx, rcx, times_4, 10000)); 432 __ movsd(Operand(rbx, rcx, times_4, 10000), xmm1);
|
/prebuilts/go/darwin-x86/pkg/bootstrap/src/bootstrap/compile/internal/x86/ |
peep.go | 224 // MOVSD removal. 225 // We never use packed registers, so a MOVSD between registers
|
/prebuilts/go/darwin-x86/src/cmd/compile/internal/x86/ |
peep.go | 221 // MOVSD removal. 222 // We never use packed registers, so a MOVSD between registers
|
/prebuilts/go/linux-x86/pkg/bootstrap/src/bootstrap/compile/internal/x86/ |
peep.go | 224 // MOVSD removal. 225 // We never use packed registers, so a MOVSD between registers
|
/prebuilts/go/linux-x86/src/cmd/compile/internal/x86/ |
peep.go | 221 // MOVSD removal. 222 // We never use packed registers, so a MOVSD between registers
|
/art/compiler/utils/x86/ |
assembler_x86_test.cc | 142 "movsd 0(%esp), %xmm0\n"
|
/art/runtime/arch/x86_64/ |
quick_entrypoints_x86_64.S | 413 movsd (%r11), REG_VAR(xmm_reg) 541 movsd %xmm0, (%r8) // Store the double floating point result. 634 movsd %xmm0, (%r8) // Store the double floating point result. [all...] |