/external/v8/src/x64/ |
code-stubs-x64.cc | 69 __ movq(rdx, Operand(rsp, 1 * kPointerSize)); 77 __ movq(rcx, Operand(rsi, Context::SlotOffset(Context::GLOBAL_INDEX))); 78 __ movq(rcx, FieldOperand(rcx, GlobalObject::kGlobalContextOffset)); 79 __ movq(rcx, Operand(rcx, Context::SlotOffset(map_index))); 80 __ movq(FieldOperand(rax, JSObject::kMapOffset), rcx); 87 __ movq(FieldOperand(rax, JSObject::kPropertiesOffset), rbx); 88 __ movq(FieldOperand(rax, JSObject::kElementsOffset), rbx); 89 __ movq(FieldOperand(rax, JSFunction::kPrototypeOrInitialMapOffset), rcx); 90 __ movq(FieldOperand(rax, JSFunction::kSharedFunctionInfoOffset), rdx); 91 __ movq(FieldOperand(rax, JSFunction::kContextOffset), rsi) [all...] |
macro-assembler-x64.cc | 73 movq(scratch, target); 83 movq(destination, Operand(kRootRegister, static_cast<int32_t>(delta))); 91 movq(kScratchRegister, source); 92 movq(destination, Operand(kScratchRegister, 0)); 102 movq(Operand(kRootRegister, static_cast<int32_t>(delta)), source); 110 movq(kScratchRegister, destination); 111 movq(Operand(kScratchRegister, 0), source); 127 movq(destination, source); 148 // Size of movq(destination, src); 155 movq(destination, Operand(kRootRegister [all...] |
regexp-macro-assembler-x64.cc | 277 __ movq(rax, combined_chars, RelocInfo::NONE); 297 __ movq(rax, combined_chars, RelocInfo::NONE); 332 __ movq(rdx, register_location(start_reg)); // Offset of start of capture 333 __ movq(rbx, register_location(start_reg + 1)); // Offset of end of capture 393 __ movq(rdi, r11); 419 __ movq(r8, rbx); 428 __ movq(rsi, rax); 430 __ movq(rdx, rbx); 468 __ movq(rdx, register_location(start_reg)); 469 __ movq(rax, register_location(start_reg + 1)) [all...] |
lithium-gap-resolver-x64.cc | 175 __ movq(dst, src); 179 __ movq(dst, src); 186 __ movq(dst, src); 190 __ movq(kScratchRegister, src); 191 __ movq(dst, kScratchRegister); 209 __ movq(dst, Immediate(cgen_->ToInteger32(constant_source))); 212 __ movq(dst, kScratchRegister); 260 __ movq(kScratchRegister, mem); 261 __ movq(mem, reg); 262 __ movq(reg, kScratchRegister) [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
subpixel_ssse3.asm | 74 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 76 movq xmm2, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 104 movq MMWORD Ptr [rdi], xmm0 133 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 135 movq xmm1, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 159 movq MMWORD Ptr [rdi], xmm0 211 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 213 movq xmm3, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 224 movq xmm3, MMWORD PTR [rsi + 6] 227 movq xmm7, MMWORD PTR [rsi + 11 [all...] |
subpixel_mmx.asm | 43 movq mm1, [rdx + 16] ; do both the negative taps first!!! 44 movq mm2, [rdx + 32] ; 45 movq mm6, [rdx + 48] ; 46 movq mm7, [rdx + 64] ; 55 movq mm3, [rsi-2] ; mm3 = p-2..p5 56 movq mm4, mm3 ; mm4 = p-2..p5 61 movq mm5, mm4 ; mm5 = p-2..p5 66 movq mm4, mm5 ; mm4 = p-2..p5; 72 movq mm5, mm4 ; mm5 = p-2..p5 93 movq [rdi], mm3 ; store the results in the destinatio [all...] |
sad_sse3.asm | 212 movq mm0, QWORD PTR [%2] 213 movq mm5, QWORD PTR [%3] 214 movq mm6, QWORD PTR [%3+1] 215 movq mm7, QWORD PTR [%3+2] 221 movq mm0, QWORD PTR [%2] 222 movq mm1, QWORD PTR [%3] 223 movq mm2, QWORD PTR [%3+1] 224 movq mm3, QWORD PTR [%3+2] 234 movq mm0, QWORD PTR [%2+%4] 235 movq mm1, QWORD PTR [%3+%5 [all...] |
postproc_mmx.asm | 60 movq mm1, QWORD ptr[rdi] ; first row 64 movq [rdi], mm1 73 movq mm1, QWORD ptr[rdi] ; first row 77 movq [rdi], mm1 102 movq mm2, mm1 ; 126 movq mm4, mm2 135 movq mm2, mm1 144 movq mm3, mm6 148 movq mm1, mm5 150 movq mm4, mm [all...] |
variance_impl_mmx.asm | 31 movq mm0, [rax] 32 movq mm1, [rax+8] 33 movq mm2, [rax+16] 34 movq mm3, [rax+24] 48 movq QWORD PTR [rsp], mm4 97 movq mm0, [rax] ; Copy eight bytes to mm0 98 movq mm1, [rbx] ; Copy eight bytes to mm1 99 movq mm2, mm0 ; Take copies 100 movq mm3, mm1 ; Take copies 116 movq mm1, [rbx] ; Copy eight bytes to mm [all...] |
idctllm_sse2.asm | 59 movq xmm0, [rax] 60 movq xmm1, [rax+rdx] 61 movq xmm2, [rax+2*rdx] 62 movq xmm3, [rax+rcx] 83 movq [rax], xmm0 84 movq [rax + rdx], xmm1 88 movq [rax], xmm2 89 movq [rax + rdx], xmm3 312 movq xmm4, [rdi] 313 movq xmm5, [rdi+rdx [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-memcpy-atom.S | [all...] |
/external/chromium_org/v8/src/x64/ |
macro-assembler-x64.cc | 82 movq(scratch, target); 92 movq(destination, Operand(kRootRegister, static_cast<int32_t>(delta))); 100 movq(kScratchRegister, source); 101 movq(destination, Operand(kScratchRegister, 0)); 111 movq(Operand(kRootRegister, static_cast<int32_t>(delta)), source); 119 movq(kScratchRegister, destination); 120 movq(Operand(kScratchRegister, 0), source); 136 movq(destination, source); 157 // Size of movq(destination, src); 166 movq(kScratchRegister, BitCast<int64_t>(kZapValue), RelocInfo::NONE64) [all...] |
regexp-macro-assembler-x64.cc | 244 __ movq(rdx, register_location(start_reg)); // Offset of start of capture 245 __ movq(rbx, register_location(start_reg + 1)); // Offset of end of capture 318 __ movq(rdi, r11); 344 __ movq(r8, rbx); 353 __ movq(rsi, rax); 355 __ movq(rdx, rbx); 393 __ movq(rdx, register_location(start_reg)); 394 __ movq(rax, register_location(start_reg + 1)); 444 __ movq(rdi, rbx); 525 __ movq(rbx, current_character()) [all...] |
lithium-gap-resolver-x64.cc | 175 __ movq(dst, src); 179 __ movq(dst, src); 186 __ movq(dst, src); 190 __ movq(kScratchRegister, src); 191 __ movq(dst, kScratchRegister); 212 __ movq(kScratchRegister, int_val, RelocInfo::NONE64); 213 __ movq(dst, kScratchRegister); 223 __ movq(dst, Immediate(cgen_->ToInteger32(constant_source))); 226 __ movq(dst, kScratchRegister); 274 __ movq(kScratchRegister, mem) [all...] |
codegen-x64.cc | 80 __ movq(rbx, xmm0); 87 __ movq(xmm0, rbx); 216 __ movq(rcx, kNaNValue, RelocInfo::NONE64); 217 __ movq(Operand(rsp, kPointerSize), rcx); 271 __ movq(FieldOperand(rdx, HeapObject::kMapOffset), rbx); 301 __ movq(r8, FieldOperand(rdx, JSObject::kElementsOffset)); 317 __ movq(r14, r8); // Destination array equals source array. 324 __ movq(FieldOperand(r14, HeapObject::kMapOffset), rdi); 328 __ movq(FieldOperand(rdx, HeapObject::kMapOffset), rbx); 343 __ movq(r15, BitCast<int64_t, uint64_t>(kHoleNanInt64), RelocInfo::NONE64) [all...] |
stub-cache-x64.cc | 77 __ movq(kScratchRegister, 84 __ movq(kScratchRegister, 125 __ movq(r0, FieldOperand(receiver, HeapObject::kMapOffset)); 141 __ movq(properties, FieldOperand(receiver, JSObject::kPropertiesOffset)); 230 __ movq(prototype, 233 __ movq(prototype, 236 __ movq(prototype, Operand(prototype, Context::SlotOffset(index))); 238 __ movq(prototype, 241 __ movq(prototype, FieldOperand(prototype, Map::kPrototypeOffset)); 262 __ movq(prototype, FieldOperand(prototype, Map::kPrototypeOffset)) [all...] |
/external/chromium_org/third_party/yasm/source/patched-yasm/modules/dbgfmts/dwarf2/tests/pass64/ |
dwarf64_leb128.asm | 163 movq %rax, conv_bv(%rip) 170 movq %rax, result(%rip) 177 movq %rax, spare(%rip) 184 movq %rax, op1static(%rip) 190 movq %rax, op2static(%rip) 193 movq %rax, from_dec_data(%rip) 211 movq from_dec_data(%rip), %rdi 214 movq op2static(%rip), %rdi 217 movq op1static(%rip), %rdi 220 movq spare(%rip), %rd [all...] |
/external/libvpx/libvpx/vpx_scale/win32/ |
scaleopt.c | 93 movq mm5, const54_1; 96 movq mm6, const54_2; 98 movq mm4, round_values; 102 movq mm0, QWORD PTR [esi]; 104 movq mm1, mm0; 165 movq mm3, mm2 174 movq mm5, mm4 241 movq mm5, const53_1; 244 movq mm6, const53_2; 246 movq mm4, round_values [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_subpixel_8t_ssse3.asm | 116 movq xmm5, rcx 148 movq xmm0, [rsi] ;A 149 movq xmm1, [rsi + rdx] ;B 150 movq xmm2, [rsi + rdx * 2] ;C 151 movq xmm3, [rax + rdx * 2] ;D 152 movq xmm4, [rsi + rdx * 4] ;E 153 movq xmm5, [rax + rdx * 4] ;F 159 movq xmm6, [rsi + rbx] ;G 160 movq xmm7, [rax + rbx] ;H 179 movq xmm1, [rdi [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_sad_sse3.asm | 131 movq mm0, QWORD PTR [%2] 132 movq mm5, QWORD PTR [%3] 133 movq mm6, QWORD PTR [%3+1] 134 movq mm7, QWORD PTR [%3+2] 140 movq mm0, QWORD PTR [%2] 141 movq mm1, QWORD PTR [%3] 142 movq mm2, QWORD PTR [%3+1] 143 movq mm3, QWORD PTR [%3+2] 153 movq mm0, QWORD PTR [%2+%4] 154 movq mm1, QWORD PTR [%3+%5 [all...] |
vp9_variance_impl_sse2.asm | 59 movq rax,xmm4 246 movq xmm1, QWORD PTR [rsi] 247 movq xmm2, QWORD PTR [rdi] 257 movq xmm2, QWORD PTR[rsi + rax] 258 movq xmm3, QWORD PTR[rdi + rdx] 270 movq xmm2, QWORD PTR[rsi + rax * 2] 271 movq xmm3, QWORD PTR[rdi + rdx * 2] 285 movq xmm2, QWORD PTR[rsi + rax] 286 movq xmm3, QWORD PTR[rdi + rdx] 297 movq xmm2, QWORD PTR[rsi + rax *2 [all...] |
/external/valgrind/main/coregrind/ |
m_cpuid.S | 64 movq $1, %rax 119 movq %rsp, %rbp 122 movq %rcx, %rdi 124 movq %rdx, %rsi 149 movq %rbp, %rsp
|
/external/chromium_org/third_party/openssl/openssl/crypto/bn/asm/ |
x86_64-mont5.pl | 105 movq 0(%rax,%r10,8),%xmm4 # set of masks denoting which 106 movq 8(%rax,%r10,8),%xmm5 # cache line contains element 107 movq 16(%rax,%r10,8),%xmm6 # denoted by 7th argument 108 movq 24(%rax,%r10,8),%xmm7 110 movq `0*$STRIDE/4-96`($bp),%xmm0 111 movq `1*$STRIDE/4-96`($bp),%xmm1 113 movq `2*$STRIDE/4-96`($bp),%xmm2 115 movq `3*$STRIDE/4-96`($bp),%xmm3 123 movq %xmm0,$m0 # m0=bp[0] 131 movq `0*$STRIDE/4-96`($bp),%xmm [all...] |
/external/openssl/crypto/bn/asm/ |
x86_64-mont5.pl | 105 movq 0(%rax,%r10,8),%xmm4 # set of masks denoting which 106 movq 8(%rax,%r10,8),%xmm5 # cache line contains element 107 movq 16(%rax,%r10,8),%xmm6 # denoted by 7th argument 108 movq 24(%rax,%r10,8),%xmm7 110 movq `0*$STRIDE/4-96`($bp),%xmm0 111 movq `1*$STRIDE/4-96`($bp),%xmm1 113 movq `2*$STRIDE/4-96`($bp),%xmm2 115 movq `3*$STRIDE/4-96`($bp),%xmm3 123 movq %xmm0,$m0 # m0=bp[0] 131 movq `0*$STRIDE/4-96`($bp),%xmm [all...] |
/external/chromium_org/v8/test/cctest/ |
test-assembler-x64.cc | 103 __ movq(rax, arg2); 129 __ movq(rbp, rsp); 159 __ movq(rax, arg2); 183 __ movq(rax, arg2); 185 __ movq(rax, rdx); 212 __ movq(rbp, rsp); 219 __ movq(rax, Operand(rbp, -3 * kStackElementSize)); 248 __ movq(rbp, rsp); 249 __ movq(rax, arg1); 252 __ movq(rax, arg2) [all...] |