Lines Matching full:next
11 ; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax
12 ; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx
13 ; X32-SSE-NEXT: cvtps2pd (%ecx), %xmm0
14 ; X32-SSE-NEXT: movups %xmm0, (%eax)
15 ; X32-SSE-NEXT: retl
19 ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax
20 ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx
21 ; X32-AVX-NEXT: vcvtps2pd (%ecx), %xmm0
22 ; X32-AVX-NEXT: vmovups %xmm0, (%eax)
23 ; X32-AVX-NEXT: retl
27 ; X64-SSE-NEXT: cvtps2pd (%rdi), %xmm0
28 ; X64-SSE-NEXT: movups %xmm0, (%rsi)
29 ; X64-SSE-NEXT: retq
33 ; X64-AVX-NEXT: vcvtps2pd (%rdi), %xmm0
34 ; X64-AVX-NEXT: vmovups %xmm0, (%rsi)
35 ; X64-AVX-NEXT: retq
46 ; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax
47 ; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx
48 ; X32-SSE-NEXT: cvtps2pd (%ecx), %xmm0
49 ; X32-SSE-NEXT: cvtps2pd 8(%ecx), %xmm1
50 ; X32-SSE-NEXT: movups %xmm1, 16(%eax)
51 ; X32-SSE-NEXT: movups %xmm0, (%eax)
52 ; X32-SSE-NEXT: retl
56 ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax
57 ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx
58 ; X32-AVX-NEXT: vcvtps2pd (%ecx), %ymm0
59 ; X32-AVX-NEXT: vmovups %ymm0, (%eax)
60 ; X32-AVX-NEXT: vzeroupper
61 ; X32-AVX-NEXT: retl
65 ; X64-SSE-NEXT: cvtps2pd (%rdi), %xmm0
66 ; X64-SSE-NEXT: cvtps2pd 8(%rdi), %xmm1
67 ; X64-SSE-NEXT: movups %xmm1, 16(%rsi)
68 ; X64-SSE-NEXT: movups %xmm0, (%rsi)
69 ; X64-SSE-NEXT: retq
73 ; X64-AVX-NEXT: vcvtps2pd (%rdi), %ymm0
74 ; X64-AVX-NEXT: vmovups %ymm0, (%rsi)
75 ; X64-AVX-NEXT: vzeroupper
76 ; X64-AVX-NEXT: retq
87 ; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax
88 ; X32-SSE-NEXT: movl {{[0-9]+}}(%esp), %ecx
89 ; X32-SSE-NEXT: cvtps2pd (%ecx), %xmm0
90 ; X32-SSE-NEXT: cvtps2pd 8(%ecx), %xmm1
91 ; X32-SSE-NEXT: cvtps2pd 16(%ecx), %xmm2
92 ; X32-SSE-NEXT: cvtps2pd 24(%ecx), %xmm3
93 ; X32-SSE-NEXT: movups %xmm3, 48(%eax)
94 ; X32-SSE-NEXT: movups %xmm2, 32(%eax)
95 ; X32-SSE-NEXT: movups %xmm1, 16(%eax)
96 ; X32-SSE-NEXT: movups %xmm0, (%eax)
97 ; X32-SSE-NEXT: retl
101 ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax
102 ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %ecx
103 ; X32-AVX-NEXT: vcvtps2pd (%ecx), %ymm0
104 ; X32-AVX-NEXT: vcvtps2pd 16(%ecx), %ymm1
105 ; X32-AVX-NEXT: vmovups %ymm1, 32(%eax)
106 ; X32-AVX-NEXT: vmovups %ymm0, (%eax)
107 ; X32-AVX-NEXT: vzeroupper
108 ; X32-AVX-NEXT: retl
112 ; X64-SSE-NEXT: cvtps2pd (%rdi), %xmm0
113 ; X64-SSE-NEXT: cvtps2pd 8(%rdi), %xmm1
114 ; X64-SSE-NEXT: cvtps2pd 16(%rdi), %xmm2
115 ; X64-SSE-NEXT: cvtps2pd 24(%rdi), %xmm3
116 ; X64-SSE-NEXT: movups %xmm3, 48(%rsi)
117 ; X64-SSE-NEXT: movups %xmm2, 32(%rsi)
118 ; X64-SSE-NEXT: movups %xmm1, 16(%rsi)
119 ; X64-SSE-NEXT: movups %xmm0, (%rsi)
120 ; X64-SSE-NEXT: retq
124 ; X64-AVX-NEXT: vcvtps2pd (%rdi), %ymm0
125 ; X64-AVX-NEXT: vcvtps2pd 16(%rdi), %ymm1
126 ; X64-AVX-NEXT: vmovups %ymm1, 32(%rsi)
127 ; X64-AVX-NEXT: vmovups %ymm0, (%rsi)
128 ; X64-AVX-NEXT: vzeroupper
129 ; X64-AVX-NEXT: retq
140 ; X32-SSE-NEXT: movaps {{.*#+}} xmm0 = [1.000000e+00,-2.000000e+00]
141 ; X32-SSE-NEXT: retl
145 ; X32-AVX-NEXT: vmovaps {{.*#+}} xmm0 = [1.000000e+00,-2.000000e+00]
146 ; X32-AVX-NEXT: retl
150 ; X64-SSE-NEXT: movaps {{.*#+}} xmm0 = [1.000000e+00,-2.000000e+00]
151 ; X64-SSE-NEXT: retq
155 ; X64-AVX-NEXT: vmovaps {{.*#+}} xmm0 = [1.000000e+00,-2.000000e+00]
156 ; X64-AVX-NEXT: retq