/dalvik/vm/mterp/x86/ |
OP_ADD_DOUBLE.S | 10 addsd %xmm1, %xmm0
|
OP_ADD_DOUBLE_2ADDR.S | 11 addsd %xmm1, %xmm0 # %xmm0<- vA op vB
|
/external/llvm/test/CodeGen/X86/ |
negative-sin.ll | 2 ; CHECK-NOT: {{addsd|subsd|xor}}
|
sincos-opt.ll | 35 ; GNU_SINCOS: addsd 8(%rsp), %xmm0 39 ; OSX_SINCOS: addsd %xmm1, %xmm0
|
pr3154.ll | 32 call void asm sideeffect "movsd $0, %xmm7 \0A\09movapd ff_pd_1, %xmm6 \0A\09movapd ff_pd_2, %xmm5 \0A\09movlhps %xmm7, %xmm7 \0A\09subpd %xmm5, %xmm7 \0A\09addsd %xmm6, %xmm7 \0A\09", "*m,~{dirflag},~{fpsr},~{flags}"(double* %c) nounwind 84 %asmtmp32 = call i32 asm sideeffect "movsd ff_pd_1, %xmm0 \0A\09movsd ff_pd_1, %xmm1 \0A\09movsd ff_pd_1, %xmm2 \0A\091: \0A\09movapd ($4,$0), %xmm3 \0A\09movupd -8($5,$0), %xmm4 \0A\09movapd ($5,$0), %xmm5 \0A\09mulpd %xmm3, %xmm4 \0A\09mulpd %xmm3, %xmm5 \0A\09mulpd -16($5,$0), %xmm3 \0A\09addpd %xmm4, %xmm1 \0A\09addpd %xmm5, %xmm0 \0A\09addpd %xmm3, %xmm2 \0A\09add $$16, $0 \0A\09jl 1b \0A\09movhlps %xmm0, %xmm3 \0A\09movhlps %xmm1, %xmm4 \0A\09movhlps %xmm2, %xmm5 \0A\09addsd %xmm3, %xmm0 \0A\09addsd %xmm4, %xmm1 \0A\09addsd %xmm5, %xmm2 \0A\09movsd %xmm0, $1 \0A\09movsd %xmm1, $2 \0A\09movsd %xmm2, $3 \0A\09", "=&r,=*m,=*m,=*m,r,r,0,~{dirflag},~{fpsr},~{flags}"(double* %28, double* %29, double* %30, double* %21, double* %27, i32 %22) nounwind ; <i32> [#uses=0] 92 %asmtmp34 = call i32 asm sideeffect "movsd ff_pd_1, %xmm0 \0A\09movsd ff_pd_1, %xmm1 \0A\091: \0A\09movapd ($3,$0), %xmm3 \0A\09movupd -8($4,$0), %xmm4 \0A\09mulpd %xmm3, %xmm4 \0A\09mulpd ($4,$0), %xmm3 \0A\09addpd %xmm4, %xmm1 \0A\09addpd %xmm3, %xmm0 \0A\09add $$16, $0 \0A\09jl 1b \0A\09movhlps %xmm0, %xmm3 \0A\09movhlps %xmm1, %xmm4 \0A\09addsd %xmm3, %xmm0 \0A\09addsd %xmm4, %xmm1 \0A\09movsd %xmm0, $1 \0A\09movsd %xmm1, $2 \0A\09", "=&r,=*m,=*m,r,r,0,~{dirflag},~{fpsr},~{flags}"(double* %32, double* %33, double* %21, double* %31, i32 %22) nounwind ; <i32> [#uses=0]
|
/external/compiler-rt/lib/i386/ |
floatdidf.S | 30 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
floatundixf.S | 30 addsd REL_ADDR(twop84_plus_twop52_neg), %xmm0 // hi - 2^52 (no rounding occurs)
|
floatundidf.S | 41 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
/external/compiler-rt/lib/x86_64/ |
floatundidf.S | 40 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/x86/tests/ |
ssewidth.asm | 8 addsd xmm1, xmm2 label 9 addsd xmm1, qword [rbx] label
|
avx.asm | 71 addsd xmm1, xmm2 label 72 addsd xmm1, [rax] label 73 addsd xmm1, qword [rax] label [all...] |
/external/valgrind/main/memcheck/tests/amd64/ |
sse_memory.stdout.exp | [all...] |
sse_memory.c | 255 TEST_INSN( &AllMask, SD,addsd)
|
/external/valgrind/main/memcheck/tests/x86/ |
sse2_memory.stdout.exp | 11 r addsd 77eca2665a8c4256d3eaa2736713076f 3f55e1b30e241fff59f1ebe418d2ba28 3f55e1b30e241fff59f1ebe418d2ba28 12 r addsd bc68bf9dda3685ee6fcaf40c7feb0484 3d83a1c16d0e812c2aea572c0f856fff bc68bf9dda3685ee2aea572c0f856fff 13 r addsd 0ffd44c56f5bf82e3de842925d012f37 6b0c886e9a6f2728ff55ac7f3661970c 0ffd44c56f5bf82eff55ac7f3661970c 14 r addsd f4148b8c668ffa2bd27f73015eb678f0 6dfc0d09037a92aa8d0bf37d096a4357 d008a342668ffa2b8d0bf37d096a4357 15 r addsd 50f94b806c444cdc23a9d13a3e4f30f5 473c0ac17342a347e9c893a36480e4c5 50f94b806c444cdce9c893a36480e4c5 16 m addsd 13c07bfc401df2e6 68d59e108ccd1018845f04fa19ed086d 13c07bfc401df2e6845f04fa19ed086d 17 m addsd 389f65c305c2a505 65e87126d6ae84b9dfb3fc8a335bc495 65e87126d6ae84b9dfb3fc8a335bc495 18 m addsd b42ae1878fdd1cc5 edfa5b770c3e56bc56e2e1f0becc5f8b b42ae1878fdd1cc556e2e1f0becc5f8b 19 m addsd f5bdd1fa4c5ecf22 368e461ae51286040ec83a8a3e8e56d7 f5bdd1fa4c5ecf220ec83a8a3e8e56d7 20 m addsd 91e3aaa730432cd5 9c56b0049bcf0832122de4537ebadd80 91e3aaa730432cd5122de4537ebadd8 [all...] |
/art/compiler/utils/x86/ |
assembler_x86.h | 281 void addsd(XmmRegister dst, XmmRegister src); 282 void addsd(XmmRegister dst, const Address& src);
|
assembler_x86.cc | 427 void X86Assembler::addsd(XmmRegister dst, XmmRegister src) { function in class:art::x86::X86Assembler 436 void X86Assembler::addsd(XmmRegister dst, const Address& src) { function in class:art::x86::X86Assembler [all...] |
/external/valgrind/main/none/tests/amd64/ |
redundantRexW.c | 191 /* addsd mem, reg f2 48 0f 58 27 rex.W addsd (%rdi),%xmm4 */ 205 after_test( "rex.W addsd (%rdi),%xmm4", regs, mem );
|
/external/llvm/lib/Target/X86/ |
README.txt | 1178 addsd %xmm1, %xmm0 1185 addsd LCPI1_24, %xmm4 1187 addsd LCPI1_25, %xmm4 1189 addsd LCPI1_26, %xmm4 1191 addsd LCPI1_27, %xmm4 1193 addsd LCPI1_28, %xmm4 1195 addsd %xmm1, %xmm4 1198 addsd %xmm4, %xmm1 1205 addsd %xmm0, %xmm0 1206 addsd LCPI1_22, %xmm [all...] |
/external/valgrind/main/exp-bbv/tests/amd64-linux/ |
rep_prefix.S | 24 addsd %xmm1,%xmm2
|
/external/chromium_org/v8/src/x64/ |
codegen-x64.cc | 649 __ addsd(double_scratch, result); 669 __ addsd(result, Operand(kScratchRegister, 8 * kDoubleSize));
|
/external/chromium_org/v8/test/cctest/ |
test-disasm-ia32.cc | 393 __ addsd(xmm1, xmm0);
|
test-disasm-x64.cc | 371 __ addsd(xmm1, xmm0);
|
/external/v8/test/cctest/ |
test-assembler-ia32.cc | 273 __ addsd(xmm0, xmm1);
|
test-disasm-ia32.cc | 375 __ addsd(xmm1, xmm0);
|
test-disasm-x64.cc | 358 __ addsd(xmm1, xmm0);
|