/external/webrtc/webrtc/common_audio/signal_processing/ |
dot_product_with_scale.c | 13 int32_t WebRtcSpl_DotProductWithScale(const int16_t* vector1, 22 sum += (vector1[i + 0] * vector2[i + 0]) >> scaling; 23 sum += (vector1[i + 1] * vector2[i + 1]) >> scaling; 24 sum += (vector1[i + 2] * vector2[i + 2]) >> scaling; 25 sum += (vector1[i + 3] * vector2[i + 3]) >> scaling; 28 sum += (vector1[i] * vector2[i]) >> scaling;
|
cross_correlation_neon.c | 16 const int16_t* vector1, 27 int16x8_t seq1_16x8 = vld1q_s16(vector1); 41 vector1 += 8; 48 sum_res += WEBRTC_SPL_MUL_16_16(*vector1, *vector2); 49 vector1++;
|
/external/arm-neon-tests/ |
ref_vabd.c | 43 vabd##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 65 DECL_VABD_VAR(vector1); 72 VLOAD(vector1, buffer, , int, s, 8, 8); 73 VLOAD(vector1, buffer, , int, s, 16, 4); 74 VLOAD(vector1, buffer, , int, s, 32, 2); 75 VLOAD(vector1, buffer, , uint, u, 8, 8); 76 VLOAD(vector1, buffer, , uint, u, 16, 4); 77 VLOAD(vector1, buffer, , uint, u, 32, 2); 78 VLOAD(vector1, buffer, , float, f, 32, 2); 79 VLOAD(vector1, buffer, q, int, s, 8, 16) [all...] |
ref_vabdl.c | 40 vabdl_##T2##W(VECT_VAR(vector1, T1, W, N), \ 62 DECL_VABDL_VAR_SHORT(vector1); 69 VLOAD(vector1, buffer, , int, s, 8, 8); 70 VLOAD(vector1, buffer, , int, s, 16, 4); 71 VLOAD(vector1, buffer, , int, s, 32, 2); 72 VLOAD(vector1, buffer, , uint, u, 8, 8); 73 VLOAD(vector1, buffer, , uint, u, 16, 4); 74 VLOAD(vector1, buffer, , uint, u, 32, 2);
|
ref_vaba.c | 40 vaba##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 61 DECL_VABA_VAR(vector1); 69 VLOAD(vector1, buffer, , int, s, 8, 8); 70 VLOAD(vector1, buffer, , int, s, 16, 4); 71 VLOAD(vector1, buffer, , int, s, 32, 2); 72 VLOAD(vector1, buffer, , uint, u, 8, 8); 73 VLOAD(vector1, buffer, , uint, u, 16, 4); 74 VLOAD(vector1, buffer, , uint, u, 32, 2); 75 VLOAD(vector1, buffer, q, int, s, 8, 16); 76 VLOAD(vector1, buffer, q, int, s, 16, 8) [all...] |
ref_vaddhn.c | 56 VECT_VAR(vector64, T1, W2, N) = INSN##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 66 DECL_VARIABLE_128BITS_VARIANTS(vector1); 71 /* Fill input vector1 and vector2 with arbitrary values */ 72 VDUP(vector1, q, int, s, 16, 8, 50*(UINT8_MAX+1)); 73 VDUP(vector1, q, int, s, 32, 4, 50*(UINT16_MAX+1)); 74 VDUP(vector1, q, int, s, 64, 2, 24*((uint64_t)UINT32_MAX+1)); 75 VDUP(vector1, q, uint, u, 16, 8, 3*(UINT8_MAX+1)); 76 VDUP(vector1, q, uint, u, 32, 4, 55*(UINT16_MAX+1)); 77 VDUP(vector1, q, uint, u, 64, 2, 3*((uint64_t)UINT32_MAX+1));
|
ref_vmul.c | 46 DECL_VARIABLE(vector1, T, W, N); \ 50 /* vector_res = OP(vector1, vector2), then store the result. */ 53 INSN##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 82 VLOAD(vector1, buffer, , int, s, 8, 8); 83 VLOAD(vector1, buffer, , int, s, 16, 4); 84 VLOAD(vector1, buffer, , int, s, 32, 2); 85 VLOAD(vector1, buffer, , uint, u, 8, 8); 86 VLOAD(vector1, buffer, , uint, u, 16, 4); 87 VLOAD(vector1, buffer, , uint, u, 32, 2); 88 VLOAD(vector1, buffer, , poly, p, 8, 8) [all...] |
ref_vmla.c | 46 DECL_VARIABLE(vector1, T, W, N); \ 55 INSN##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 83 VLOAD(vector1, buffer, , int, s, 8, 8); 84 VLOAD(vector1, buffer, , int, s, 16, 4); 85 VLOAD(vector1, buffer, , int, s, 32, 2); 86 VLOAD(vector1, buffer, , uint, u, 8, 8); 87 VLOAD(vector1, buffer, , uint, u, 16, 4); 88 VLOAD(vector1, buffer, , uint, u, 32, 2); 89 VLOAD(vector1, buffer, , float, f, 32, 2); 90 VLOAD(vector1, buffer, q, int, s, 8, 16) [all...] |
ref_vabal.c | 40 vabal_##T2##W(VECT_VAR(vector1, T1, W2, N), \ 63 DECL_VABAL_VAR_LONG(vector1); 71 VLOAD(vector1, buffer, q, int, s, 16, 8); 72 VLOAD(vector1, buffer, q, int, s, 32, 4); 73 VLOAD(vector1, buffer, q, int, s, 64, 2); 74 VLOAD(vector1, buffer, q, uint, u, 16, 8); 75 VLOAD(vector1, buffer, q, uint, u, 32, 4); 76 VLOAD(vector1, buffer, q, uint, u, 64, 2);
|
ref_vqadd.c | 40 DECL_VARIABLE_ALL_VARIANTS(vector1); 44 /* Initialize input "vector1" from "buffer" */ 45 TEST_MACRO_ALL_VARIANTS_2_5(VLOAD, vector1, buffer); 87 VDUP(vector1, q, int, s, 64, 2, 0x4000000000000000LL); 127 VDUP(vector1, , uint, u, 8, 8, 0xF0); 128 VDUP(vector1, , uint, u, 16, 4, 0xFFF0); 129 VDUP(vector1, , uint, u, 32, 2, 0xFFFFFFF0); 130 VDUP(vector1, q, uint, u, 8, 16, 0xF0); 131 VDUP(vector1, q, uint, u, 16, 8, 0xFFF0); 132 VDUP(vector1, q, uint, u, 32, 4, 0xFFFFFFF0) [all...] |
ref_vqsub.c | 40 DECL_VARIABLE_ALL_VARIANTS(vector1); 44 /* Initialize input "vector1" from "buffer" */ 45 TEST_MACRO_ALL_VARIANTS_2_5(VLOAD, vector1, buffer); 90 in vector1 */ 91 VDUP(vector1, q, int, s, 64, 2, 0x3fffffffffffffffLL); 133 VDUP(vector1, , uint, u, 8, 8, 0x10); 134 VDUP(vector1, , uint, u, 16, 4, 0x10); 135 VDUP(vector1, , uint, u, 32, 2, 0x10); 136 VDUP(vector1, q, uint, u, 8, 16, 0x10); 137 VDUP(vector1, q, uint, u, 16, 8, 0x10) [all...] |
ref_vext.c | 37 /* vector_res = vext(vector1,vector2,offset), then store the result. */ 40 vext##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 47 DECL_VARIABLE_ALL_VARIANTS(vector1); 53 TEST_MACRO_ALL_VARIANTS_2_5(VLOAD, vector1, buffer); 54 VLOAD(vector1, buffer, , float, f, 32, 2); 55 VLOAD(vector1, buffer, q, float, f, 32, 4);
|
ref_vuzp.c | 57 INSN##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 71 DECL_VARIABLE_ALL_VARIANTS(vector1); 98 TEST_MACRO_ALL_VARIANTS_2_5(VLOAD, vector1, buffer); 99 VLOAD(vector1, buffer, , float, f, 32, 2); 100 VLOAD(vector1, buffer, q, float, f, 32, 4);
|
ref_v_binary_sat_op.c | 41 /* vector_res = OP(vector1,vector2), then store the result. */ 46 INSN##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 58 DECL_VARIABLE_ALL_VARIANTS(vector1); 64 /* Initialize input "vector1" from "buffer" */ 65 TEST_MACRO_ALL_VARIANTS_2_5(VLOAD, vector1, buffer);
|
/system/core/libutils/tests/ |
Vector_test.cpp | 120 Vector<int> vector1; local 121 vector1.add(1); 122 vector1.add(2); 123 vector1.add(3); 124 vector1.add(4); 126 Vector<int> vector2 = vector1; 127 ASSERT_EQ(vector1.array(), vector2.array()); 130 ASSERT_NE(vector1.editArray(), vector2.editArray()); 133 ASSERT_EQ(vector1.size(), vector2.size()); 134 for (size_t i = 0; i < vector1.size(); ++i) [all...] |
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/ |
tensor_utils_impl.h | 44 void PortableVectorVectorCwiseProduct(const float* vector1, 47 void NeonVectorVectorCwiseProduct(const float* vector1, const float* vector2, 52 void PortableVectorVectorCwiseProductAccumulate(const float* vector1, 55 void NeonVectorVectorCwiseProductAccumulate(const float* vector1, 60 float PortableVectorVectorDotProduct(const float* vector1, const float* vector2, 62 float NeonVectorVectorDotProduct(const float* vector1, const float* vector2, 66 void PortableBatchVectorBatchVectorDotProduct(const float* vector1, 70 void NeonBatchVectorBatchVectorDotProduct(const float* vector1,
|
neon_tensor_utils.h | 35 void VectorVectorCwiseProduct(const float* vector1, const float* vector2, 37 NEON_OR_PORTABLE(VectorVectorCwiseProduct, vector1, vector2, v_size, result); 40 void VectorVectorCwiseProductAccumulate(const float* vector1, 43 NEON_OR_PORTABLE(VectorVectorCwiseProductAccumulate, vector1, vector2, v_size, 54 float VectorVectorDotProduct(const float* vector1, const float* vector2, 56 return NEON_OR_PORTABLE(VectorVectorDotProduct, vector1, vector2, v_size); 59 void BatchVectorBatchVectorDotProduct(const float* vector1, 63 NEON_OR_PORTABLE(BatchVectorBatchVectorDotProduct, vector1, vector2, v_size,
|
neon_tensor_utils.cc | 68 // Load 4 float values from vector1 and vector2 and accumulator. 96 // Load 4 float values from vector1 and vector2 and accumulator. 116 void NeonVectorVectorCwiseProduct(const float* vector1, const float* vector2, 124 // Load 4 float values from vector1 and vector2. 125 float32x4_t v1_f32x4 = vld1q_f32(vector1 + v); 133 result[v] = vector1[v] * vector2[v]; 137 void NeonVectorVectorCwiseProductAccumulate(const float* vector1, 146 // Load 4 float values from vector1 and vector2 and accumulator. 147 float32x4_t v1_f32x4 = vld1q_f32(vector1 + v); 156 result[v] += vector1[v] * vector2[v] [all...] |
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/ |
portable_tensor_utils.h | 37 void PortableVectorVectorCwiseProduct(const float* vector1, 43 void PortableVectorVectorCwiseProductAccumulate(const float* vector1, 48 float PortableVectorVectorDotProduct(const float* vector1, const float* vector2, 52 void PortableBatchVectorBatchVectorDotProduct(const float* vector1, 114 void VectorVectorCwiseProduct(const float* vector1, const float* vector2, 116 PortableVectorVectorCwiseProduct(vector1, vector2, v_size, result); 119 void VectorVectorCwiseProductAccumulate(const float* vector1, 122 PortableVectorVectorCwiseProductAccumulate(vector1, vector2, v_size, result); 132 float VectorVectorDotProduct(const float* vector1, const float* vector2, 134 return PortableVectorVectorDotProduct(vector1, vector2, v_size) [all...] |
portable_tensor_utils.cc | 48 void PortableVectorVectorCwiseProduct(const float* vector1, 52 *result++ = *vector1++ * *vector2++; 56 float PortableVectorVectorDotProduct(const float* vector1, const float* vector2, 60 result += *vector1++ * *vector2++; 65 void PortableBatchVectorBatchVectorDotProduct(const float* vector1, 70 const float* vector1_ptr = vector1; 81 void PortableVectorVectorCwiseProductAccumulate(const float* vector1, 85 *result++ += *vector1++ * *vector2++;
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/ |
tensor_utils.h | 38 void VectorVectorCwiseProduct(const float* vector1, const float* vector2, 43 void VectorVectorCwiseProductAccumulate(const float* vector1, 48 float VectorVectorDotProduct(const float* vector1, const float* vector2, 52 // vector1 = [x_1_1, x_1_2, ..., x_1_vsize, 66 void BatchVectorBatchVectorDotProduct(const float* vector1,
|
/external/compiler-rt/lib/tsan/tests/unit/ |
tsan_clock_test.cc | 57 ThreadClock vector1(100); 58 vector1.tick(); 60 vector1.release(&cache, &chunked); 113 ThreadClock vector1(10); 114 vector1.tick(); 119 vector1.release(&cache, &chunked); 129 vector1.release(&cache, &chunked); 135 vector1.release(&cache, &chunked); 143 vector1.acquire(&cache, &chunked); 144 ASSERT_EQ(vector1.size(), 21U) [all...] |
/frameworks/base/libs/hwui/ |
ShadowTessellator.h | 86 static int getExtraVertexNumber(const Vector2& vector1, const Vector2& vector2, float divisor);
|
ShadowTessellator.cpp | 164 int ShadowTessellator::getExtraVertexNumber(const Vector2& vector1, const Vector2& vector2, 167 if (vector1.lengthSquared() == 0 || vector2.lengthSquared() == 0) { 176 float dotProduct = vector1.dot(vector2);
|
/frameworks/base/core/java/android/gesture/ |
GestureUtils.java | 403 * @param vector1 407 static float squaredEuclideanDistance(float[] vector1, float[] vector2) { 409 int size = vector1.length; 411 float difference = vector1[i] - vector2[i]; 420 * @param vector1 424 static float cosineDistance(float[] vector1, float[] vector2) { 426 int len = vector1.length; 428 sum += vector1[i] * vector2[i]; 436 * @param vector1 441 static float minimumCosineDistance(float[] vector1, float[] vector2, int numOrientations) [all...] |