/external/zlib/contrib/amd64/ |
amd64-match.S | 269 mov $(-MAX_MATCH_8), %rdx 275 prefetcht1 (%windowbestlen, %rdx) 276 prefetcht1 (%prev, %rdx) 280 * adjust %rdx so that it is offset to the exact byte that mismatched. 293 movdqu (%windowbestlen, %rdx), %xmm1 294 movdqu (%prev, %rdx), %xmm2 296 movdqu 16(%windowbestlen, %rdx), %xmm3 297 movdqu 16(%prev, %rdx), %xmm4 299 movdqu 32(%windowbestlen, %rdx), %xmm5 300 movdqu 32(%prev, %rdx), %xmm [all...] |
/external/openssl/crypto/bn/asm/ |
x86_64-mont.pl | 34 $bp="%rdx"; # const BN_ULONG *bp, 70 mov %rdx,$bp # $bp reassigned, remember? 81 mov %rdx,$hi0 88 adc \$0,%rdx 89 mov %rdx,$hi1 96 adc \$0,%rdx 99 mov %rdx,$hi0 104 adc \$0,%rdx 106 adc \$0,%rdx 109 mov %rdx,$hi [all...] |
/external/v8/src/x64/ |
ic-x64.cc | 270 __ CmpObjectType(rcx, JS_OBJECT_TYPE, rdx); 274 __ testb(FieldOperand(rdx, Map::kBitFieldOffset), 322 __ CmpObjectType(rax, FIRST_NONSTRING_TYPE, rdx); 331 __ testb(FieldOperand(rdx, Map::kInstanceTypeOffset), 344 __ movl(rdx, rbx); 345 __ shr(rdx, Immediate(KeyedLookupCache::kMapHashShift)); 348 __ xor_(rdx, rax); 349 __ and_(rdx, Immediate(KeyedLookupCache::kCapacityMask)); 355 __ movq(rdi, rdx); 369 __ movl(rax, Operand(kScratchRegister, rdx, times_4, 0)) 1000 __ movq(Operand(rsp, (argc + 1) * kPointerSize), rdx); local 1103 __ movq(Operand(rsp, (argc + 1) * kPointerSize), rdx); local [all...] |
stub-cache-x64.cc | 677 __ movq(rdx, Operand(rsp, (argc + 1) * kPointerSize)); 681 __ JumpIfSmi(rdx, &miss); 691 CheckPrototypes(JSObject::cast(object), rdx, holder, local 697 __ movq(rdx, FieldOperand(rdx, GlobalObject::kGlobalReceiverOffset)); 698 __ movq(Operand(rsp, (argc + 1) * kPointerSize), rdx); local 708 __ CmpObjectType(rdx, FIRST_NONSTRING_TYPE, rax); 715 rbx, rdx, name, &miss); 726 __ JumpIfSmi(rdx, &fast); 727 __ CmpObjectType(rdx, HEAP_NUMBER_TYPE, rax) 763 CheckPrototypes(JSObject::cast(object), rdx, holder, local 830 __ movq(Operand(rsp, (argc + 1) * kPointerSize), rdx); local 894 __ movq(Operand(rsp, (argc + 1) * kPointerSize), rdx); local 970 __ movq(Operand(rsp, (argc + 1) * kPointerSize), rdx); local 1195 GenerateLoadArrayLength(masm(), rcx, rdx, &miss); local 1251 GenerateLoadFunctionPrototype(masm(), rcx, rdx, rbx, &miss); local 1315 GenerateLoadStringLength(masm(), rcx, rdx, rbx, &miss); local 1392 rdx, rcx, rbx, local 1546 rdx, rcx, rbx, local [all...] |
builtins-x64.cc | 109 // -- rdx : code entry to call 172 __ call(rdx); 182 __ jmp(rdx); 297 __ GetBuiltinEntry(rdx, Builtins::CALL_NON_FUNCTION); 306 __ movq(rdx, FieldOperand(rdi, JSFunction::kSharedFunctionInfoOffset)); 308 FieldOperand(rdx, SharedFunctionInfo::kFormalParameterCountOffset)); 309 __ movq(rdx, FieldOperand(rdx, SharedFunctionInfo::kCodeOffset)); 310 __ lea(rdx, FieldOperand(rdx, Code::kHeaderSize)) 987 __ movq(Operand(rcx, 0), rdx); local 1040 __ movl(Operand(rdi, FixedArray::kLengthOffset), rdx); \/\/ and length local 1052 __ movq(Operand(rcx, 0), rdx); local [all...] |
/external/libvpx/vp8/common/x86/ |
subpixel_ssse3.asm | 47 movsxd rdx, DWORD PTR arg(5) ;table index 49 shl rdx, 4 54 add rax, rdx 68 movsxd rdx, dword ptr arg(3) ;output_pitch 70 sub rdi, rdx 84 add rdi, rdx 118 movsxd rdx, dword ptr arg(3) ;output_pitch 120 sub rdi, rdx 130 add rdi, rdx 171 movsxd rdx, DWORD PTR arg(5) ;table inde [all...] |
subpixel_sse2.asm | 46 mov rdx, arg(6) ;vp8_filter 76 pmullw xmm3, XMMWORD PTR [rdx] ; x[-2] * H[-2]; Tap 1 80 pmullw xmm4, XMMWORD PTR [rdx+16] ; x[-1] * H[-1]; Tap 2 86 pmullw xmm5, [rdx+32] ; x[ 0] * H[ 0]; Tap 3 91 pmullw xmm6, [rdx+48] ; x[ 1] * h[ 1] ; Tap 4 97 pmullw xmm7, [rdx+64] ; x[ 2] * h[ 2] ; Tap 5 100 pmullw xmm1, [rdx+80] ; x[ 3] * h[ 3] ; Tap 6 166 mov rdx, arg(6) ;vp8_filter 201 pmullw xmm3, XMMWORD PTR [rdx] ; x[-2] * H[-2]; Tap 1 205 pmullw xmm4, XMMWORD PTR [rdx+16] ; x[-1] * H[-1]; Tap [all...] |
idctllm_sse2.asm | 32 mov rdx, arg(1) ; dequant 39 movd xmm5, [rdx] 42 pinsrw xmm5, [rdx], 4 71 movsxd rdx, dword ptr arg(4) ; dst_stride 87 movq [rax + rdx], xmm1 89 lea rax, [rax + 2*rdx] 92 movq [rax + rdx], xmm3 120 mov rdx, arg(1) ; dequant 136 pmullw xmm0, [rdx] 137 pmullw xmm2, [rdx+16 [all...] |
idctllm_mmx.asm | 45 mov rdx, arg(1) ;output 172 movq [rdx], mm0 174 movq [rdx+rax], mm1 175 movq [rdx+rax*2], mm2 177 add rdx, rax 178 movq [rdx+rax*2], mm5 200 mov rdx, arg(1) ;output 208 movq [rdx], mm0 209 movq [rdx+rax], mm0 211 movq [rdx+rax*2], mm [all...] |
subpixel_mmx.asm | 40 mov rdx, arg(6) ;vp8_filter 42 movq mm1, [rdx + 16] ; do both the negative taps first!!! 43 movq mm2, [rdx + 32] ; 44 movq mm6, [rdx + 48] ; 45 movq mm7, [rdx + 64] ; 80 pmullw mm4, [rdx+80] ; mm5 *= kernel 0 modifiers 84 pmullw mm5, [rdx] ; mm5 *= kernel 5 modifiers 147 movsxd rdx, dword ptr arg(2) ;pixels_per_line 150 sub rsi, rdx 151 sub rsi, rdx [all...] |
recon_sse2.asm | 25 mov rdx, arg(1) ;q 31 paddsw xmm1, XMMWORD PTR [rdx] 38 paddsw xmm2, XMMWORD PTR [rdx+16] 45 paddsw xmm3, XMMWORD PTR [rdx+32] 52 paddsw xmm4, XMMWORD PTR [rdx+48] 77 mov rdx, arg(1) ;q 85 paddsw xmm1, XMMWORD PTR [rdx] 86 paddsw xmm5, XMMWORD PTR [rdx+16] 95 paddsw xmm2, XMMWORD PTR [rdx+32] 96 paddsw xmm6, XMMWORD PTR [rdx+48 [all...] |
loopfilter_sse2.asm | 160 mov rdx, arg(2) ; get flimit 161 movdqa xmm2, XMMWORD PTR [rdx] 172 mov rdx, arg(4) ; get thresh 173 movdqa xmm7, XMMWORD PTR [rdx] 319 mov rdx, arg(3) ;limit 320 movdqa xmm7, XMMWORD PTR [rdx] 382 mov rdx, arg(3) ;limit 383 movdqa xmm7, XMMWORD PTR [rdx] 643 mov rdx, arg(3) ;limit 644 movdqa xmm7, XMMWORD PTR [rdx] [all...] |
/external/zlib/contrib/gcc_gvmat64/ |
gvmat64.S | 68 ; register used : rax,rbx,rcx,rdx,rsi,rdi,r8,r9,r10,r11,r12
170 ; parameter 1 in rcx(deflate state s), param 2 in rdx (cur match)
176 ; rax, rcx, rdx, r8, r9, r10, and r11, which are scratch.
195 ; this clear high 32 bits of r8, which can be garbage in both r8 and rdx
419 mov rdx, 0xfffffffffffffef8 //; -(MAX_MATCH_8)
423 prefetcht1 [rsi+rdx]
424 prefetcht1 [rdi+rdx]
428 ;;; adjust rdx so that it is offset to the exact byte that mismatched.
443 mov rax, [rsi + rdx]
444 xor rax, [rdi + rdx]
[all...] |
/external/libvpx/vp8/encoder/x86/ |
subtract_mmx.asm | 30 movsxd rdx, dword ptr arg(1);src_stride; 42 movd mm0, [rsi+rdx] 50 movd mm0, [rsi+rdx*2] 57 lea rsi, [rsi+rdx*2] 62 movd mm0, [rsi+rdx] 91 movsxd rdx, dword ptr arg(3) ;stride 139 lea rsi, [rsi+rdx] 176 movsxd rdx, dword ptr arg(4) ;stride; 193 movq mm0, [rsi+rdx] 206 movq mm0, [rsi+rdx*2 [all...] |
quantize_mmx.asm | 46 mov rdx, arg(6) ;quant_ptr 47 movq mm1, [rdx] 89 movq mm5, [rdx+8] 130 movq mm5, [rdx+16] 171 movq mm5, [rdx+24] 255 xor rdx, rdx 256 sub rdx, rcx ; rdx=-rcx 261 sar rdx, 3 [all...] |
sad_sse3.asm | 41 lddqu xmm1, QWORD PTR [rdi+rdx] 42 lddqu xmm2, QWORD PTR [rdi+rdx+1] 43 lddqu xmm3, QWORD PTR [rdi+rdx+2] 46 lea rdi, [rdi+rdx*2] 82 movq mm1, QWORD PTR [rdi+rdx] 83 movq mm2, QWORD PTR [rdi+rdx+1] 84 movq mm3, QWORD PTR [rdi+rdx+2] 87 lea rdi, [rdi+rdx*2] 110 lddqu xmm5, [rdx] 121 lddqu xmm2, [rdx] [all...] |
sad_sse2.asm | 34 movsxd rdx, dword ptr arg(3) ;ref_stride 50 movq xmm5, QWORD PTR [rdi+rdx] 59 movq xmm3, QWORD PTR [rdi+rdx+8] 62 lea rdi, [rdi+rdx*2] 107 movsxd rdx, dword ptr arg(3) ;ref_stride 124 movq mm3, QWORD PTR [rdi+rdx] 130 lea rdi, [rdi+rdx*2] 170 movsxd rdx, dword ptr arg(3) ;ref_stride 187 add rdi, rdx 222 movsxd rdx, dword ptr arg(3) ;ref_strid [all...] |
sad_ssse3.asm | 41 lddqu xmm1, QWORD PTR [rdi+rdx] 42 lddqu xmm2, QWORD PTR [rdi+rdx+1] 43 lddqu xmm3, QWORD PTR [rdi+rdx+2] 46 lea rdi, [rdi+rdx*2] 96 movdqa xmm4, QWORD PTR [rdi+rdx] 97 movdqa xmm3, QWORD PTR [rdi+rdx+16] 108 lea rdi, [rdi+rdx*2] 170 mov rdx, 0xf 171 and rdx, rdi 199 movsxd rax, dword [rax + 4*rdx] ; get the 32 bit offset from the jumptabl [all...] |
/bionic/libc/kernel/arch-x86/asm/ |
ptrace.h | 58 unsigned long rdx; member in struct:pt_regs
|
/development/ndk/platforms/android-5/arch-x86/include/asm/ |
ptrace.h | 58 unsigned long rdx; member in struct:pt_regs
|
/prebuilt/ndk/android-ndk-r4/platforms/android-5/arch-x86/usr/include/asm/ |
ptrace.h | 58 unsigned long rdx; member in struct:pt_regs
|
/prebuilt/ndk/android-ndk-r4/platforms/android-8/arch-x86/usr/include/asm/ |
ptrace.h | 58 unsigned long rdx; member in struct:pt_regs
|
/external/libffi/src/x86/ |
unix64.S | 50 movq %rdx, (%rax) /* Save flags. */ 63 movq 16(%r10), %rdx 172 cmovnz %rax, %rdx 175 cmovnz %r10, %rdx 178 cmovnz %r11, %rdx 180 movq %rdx, 8(%rsi) 219 movq %rdx, 16(%rsp) 228 movq %rsp, %rdx 298 /* There are four possibilities here, %rax/%rdx, %xmm0/%rax, 300 both rdx and xmm1 with the second word. For the remaining [all...] |
/external/libvpx/vp8/decoder/x86/ |
dequantize_mmx.asm | 65 mov rdx, arg(1) ;dq 69 pmullw mm0, [rdx] 72 pmullw mm1, [rdx +8] 75 pmullw mm2, [rdx+16] 78 pmullw mm3, [rdx+24] 80 mov rdx, arg(3) ;dest 218 movd [rdx], mm0 224 movd [rdx+rdi], mm1 230 movd [rdx+rdi*2], mm2 232 add rdx, rd [all...] |
/external/v8/test/cctest/ |
test-macro-assembler-x64.cc | 54 using v8::internal::rdx; 83 // in RSI, RDI, RDX, RCX, R8, and R9, and floating point arguments in 121 __ Set(rdx, reinterpret_cast<intptr_t>(Smi::FromInt(0))); 122 __ cmpq(rcx, rdx); 169 __ Move(rdx, Smi::FromInt(y)); 170 __ movq(r9, rdx); 171 __ SmiCompare(rcx, rdx); 187 __ cmpq(rdx, r9); 191 __ SmiCompare(rdx, rcx); 277 __ Set(rdx, reinterpret_cast<intptr_t>(Smi::FromInt(0))) [all...] |