/toolchain/binutils/binutils-2.27/gas/testsuite/gas/alpha/ |
elf-usepv-1.d | 2 #name: alpha elf-usepv-1 4 .*: file format elf64-alpha.*
|
/external/skia/src/core/ |
SkBlitRow_D32.cpp | 18 int count, U8CPU alpha) { 19 SkASSERT(255 == alpha); 25 int count, U8CPU alpha) { 26 SkASSERT(alpha <= 255); 28 unsigned src_scale = SkAlpha255To256(alpha); 59 int count, U8CPU alpha) { 60 SkASSERT(alpha <= 255); 64 *dst = SkBlendARGB32(*(src++), *dst, alpha); 71 *dst = SkBlendARGB32(*(src++), *dst, alpha); 73 *dst = SkBlendARGB32(*(src++), *dst, alpha); [all...] |
SkAlphaRuns.cpp | 31 const uint8_t* alpha = fAlpha; local 34 SkASSERT(*alpha <= max); 35 alpha += *runs; 42 const uint8_t* alpha = fAlpha; local 48 SkDebugf(" %02x", *alpha); 52 alpha += n;
|
/external/skqp/src/core/ |
SkBlitRow_D32.cpp | 18 int count, U8CPU alpha) { 19 SkASSERT(255 == alpha); 25 int count, U8CPU alpha) { 26 SkASSERT(alpha <= 255); 28 unsigned src_scale = SkAlpha255To256(alpha); 59 int count, U8CPU alpha) { 60 SkASSERT(alpha <= 255); 64 *dst = SkBlendARGB32(*(src++), *dst, alpha); 71 *dst = SkBlendARGB32(*(src++), *dst, alpha); 73 *dst = SkBlendARGB32(*(src++), *dst, alpha); [all...] |
SkAlphaRuns.cpp | 31 const uint8_t* alpha = fAlpha; local 34 SkASSERT(*alpha <= max); 35 alpha += *runs; 42 const uint8_t* alpha = fAlpha; local 48 SkDebugf(" %02x", *alpha); 52 alpha += n;
|
/frameworks/base/packages/SystemUI/src/com/android/systemui/statusbar/ |
CrossFadeHelper.java | 38 .alpha(0f) 81 float alpha = Interpolators.ALPHA_OUT.getInterpolation(1.0f - fadeOutAmount); local 82 view.setAlpha(alpha); 83 updateLayerType(view, alpha); 93 private static void updateLayerType(View view, float alpha) { 94 if (view.hasOverlappingRendering() && alpha > 0.0f && alpha < 1.0f) { 112 .alpha(1f) 145 float alpha = Interpolators.ALPHA_IN.getInterpolation(fadeInAmount); local 146 view.setAlpha(alpha); [all...] |
/external/apache-commons-math/src/main/java/org/apache/commons/math/stat/inference/ |
ChiSquareTest.java | 93 * significance level <code>alpha</code>. Returns true iff the null hypothesis can be rejected 94 * with 100 * (1 - alpha) percent confidence. 108 * <li> <code> 0 < alpha < 0.5 </code> 115 * @param alpha significance level of the test 117 * 1 - alpha 121 boolean chiSquareTest(double[] expected, long[] observed, double alpha) 188 * with significance level <code>alpha</code>. Returns true iff the null hypothesis can be rejected 189 * with 100 * (1 - alpha) percent confidence. 213 * @param alpha significance level of the test 215 * 1 - alpha [all...] |
/external/robolectric/v1/src/main/java/com/xtremelabs/robolectric/shadows/ |
ShadowColor.java | 16 public static int argb(int alpha, int red, int green, int blue) { 17 return (alpha << 24) | (red << 16) | (green << 8) | blue;
|
/external/skia/src/opts/ |
SkBlitRow_opts_SSE2.h | 15 int count, U8CPU alpha); 19 int count, U8CPU alpha);
|
/external/skqp/src/opts/ |
SkBlitRow_opts_SSE2.h | 15 int count, U8CPU alpha); 19 int count, U8CPU alpha);
|
/frameworks/ml/nn/runtime/test/specs/V1_1/ |
local_response_norm_float_1_relaxed.mod.py | 21 alpha = Float32Scalar("alpha", 4.) variable 25 model = model.Operation("LOCAL_RESPONSE_NORMALIZATION", i1, radius, bias, alpha, beta).To(output)
|
local_response_norm_float_2_relaxed.mod.py | 21 alpha = Float32Scalar("alpha", 1.) variable 25 model = model.Operation("LOCAL_RESPONSE_NORMALIZATION", i1, radius, bias, alpha, beta).To(output)
|
local_response_norm_float_3_relaxed.mod.py | 21 alpha = Float32Scalar("alpha", 4.) variable 25 model = model.Operation("LOCAL_RESPONSE_NORMALIZATION", i1, radius, bias, alpha, beta).To(output)
|
local_response_norm_float_4_relaxed.mod.py | 21 alpha = Float32Scalar("alpha", 4.) variable 25 model = model.Operation("LOCAL_RESPONSE_NORMALIZATION", i1, radius, bias, alpha, beta).To(output)
|
/toolchain/binutils/binutils-2.27/ld/testsuite/ld-elf/ |
compressed1a.d | 5 #notarget: alpha-*
|
pr17550a.d | 5 #notarget: alpha-*-* cr16-*-* crx-*-* d30v-*-* dlx-*-* i960-*-* pj*-*-* 6 # Disabled on alpha because alpha has a different .set directive.
|
pr17550b.d | 5 #notarget: alpha-*-* cr16-*-* crx-*-* d30v-*-* dlx-*-* i960-*-* pj*-*-* 6 # Disabled on alpha because alpha has a different .set directive.
|
pr17550d.d | 5 #notarget: alpha-*-* cr16-*-* crx-*-* d30v-*-* dlx-*-* i960-*-* pj*-*-* 6 # Disabled on alpha because alpha has a different .set directive.
|
/external/cblas/testing/ |
c_cblas1.c | 11 void F77_caxpy(const int *N, const void *alpha, void *X, 14 cblas_caxpy(*N, alpha, X, *incX, Y, *incY); 39 void F77_cscal(const int *N, const void * *alpha, void *X, 42 cblas_cscal(*N, alpha, X, *incX); 46 void F77_csscal(const int *N, const float *alpha, void *X, 49 cblas_csscal(*N, *alpha, X, *incX);
|
c_zblas1.c | 11 void F77_zaxpy(const int *N, const void *alpha, void *X, 14 cblas_zaxpy(*N, alpha, X, *incX, Y, *incY); 39 void F77_zdscal(const int *N, const double *alpha, void *X, 42 cblas_zdscal(*N, *alpha, X, *incX); 46 void F77_zscal(const int *N, const void * *alpha, void *X, 49 cblas_zscal(*N, alpha, X, *incX);
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
axpy_simple_test.cc | 32 auto alpha = builder.ConstantR0<float>(3.1415926535); local 35 auto ax = builder.Mul(alpha, x); 45 auto alpha = builder.ConstantR0<float>(3.1415926535); local 48 auto ax = builder.Mul(alpha, x); 57 auto alpha = builder.ConstantR0<float>(3.1415926535); local 62 auto ax = builder.Mul(alpha, x);
|
/external/opencv/cvaux/include/ |
cvmat.hpp | 650 /* (A^t)*alpha */ 654 explicit _CvMAT_T_( const CvMAT* a, double alpha ); 661 double alpha; 676 /* (A^ta)*(B^tb)*alpha */ 681 double alpha, int t_abc ); 684 double alpha; 691 /* (A^ta)*(B^tb)*alpha + (C^tc)*beta */ 696 explicit _CvMAT_MUL_ADD_( const CvMAT* a, const CvMAT* b, double alpha, 700 double alpha, beta; 723 /* A*alpha + B*beta + gamma * [all...] |
/external/tensorflow/tensorflow/contrib/nn/python/ops/ |
scaled_softplus.py | 39 def scaled_softplus(x, alpha, clip=None, name=None): 40 """Returns `y = alpha * ln(1 + exp(x / alpha))` or `min(y, clip)`. 43 appropriately scaled. As `alpha` tends to 0, `scaled_softplus(x, alpha)` tends 44 to `relu(x)`. The clipping is optional. As alpha->0, scaled_softplus(x, alpha) 45 tends to relu(x), and scaled_softplus(x, alpha, clip=6) tends to relu6(x). 52 alpha: A `Tensor`, indicating the amount of smoothness. The caller 53 must ensure that `alpha > 0` [all...] |
/external/tensorflow/tensorflow/python/kernel_tests/distributions/ |
dirichlet_multinomial_test.py | 38 alpha = np.random.rand(3) 39 dist = ds.DirichletMultinomial(1., alpha) 47 alpha = np.random.rand(3, 2, 2) 49 dist = ds.DirichletMultinomial(n, alpha) 56 alpha = [[1., 2, 3]] 59 dist = ds.DirichletMultinomial(n, alpha) 64 alpha = [[1., 2, 3]] 66 dist = ds.DirichletMultinomial(1, alpha) 68 self.assertAllClose(alpha, dist.concentration.eval()) 71 alpha = [[1., 2, 3] [all...] |
/external/adhd/cras/src/dsp/ |
biquad.c | 59 double alpha = 0.25 * (0.5 + beta - gamma); local 61 double b0 = 2 * alpha; 62 double b1 = 2 * 2 * alpha; 63 double b2 = 2 * alpha; 94 double alpha = 0.25 * (0.5 + beta + gamma); local 96 double b0 = 2 * alpha; 97 double b1 = 2 * -2 * alpha; 98 double b2 = 2 * alpha; 124 double alpha = sin(w0) / (2 * Q); local 127 double b0 = alpha; 166 double alpha = 0.5 * sin(w0) * local 201 double alpha = 0.5 * sin(w0) * local 236 double alpha = sin(w0) \/ (2 * Q); local 271 double alpha = sin(w0) \/ (2 * Q); local 306 double alpha = sin(w0) \/ (2 * Q); local [all...] |