HomeSort by relevance Sort by last modified time
    Searched refs:inference (Results 1 - 25 of 43) sorted by null

1 2

  /packages/apps/Dialer/java/com/android/dialer/p13n/inference/protocol/
P13nRankerFactory.java 15 package com.android.dialer.p13n.inference.protocol;
P13nRanker.java 15 package com.android.dialer.p13n.inference.protocol;
55 * Refreshes ranking cache (pulls fresh contextual features, pre-caches inference results, etc.).
  /external/apache-commons-math/src/main/java/org/apache/commons/math/stat/inference/
OneWayAnova.java 17 package org.apache.commons.math.stat.inference;
28 * the {@link org.apache.commons.math.stat.inference.TTest}.
UnknownDistributionChiSquareTest.java 17 package org.apache.commons.math.stat.inference;
ChiSquareTest.java 17 package org.apache.commons.math.stat.inference;
TTest.java 17 package org.apache.commons.math.stat.inference;
    [all...]
OneWayAnovaImpl.java 17 package org.apache.commons.math.stat.inference;
TestUtils.java 17 package org.apache.commons.math.stat.inference;
24 * A collection of static methods to create inference test instances or to
25 * perform inference tests.
156 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticT(double[], double[])
164 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticT(org.apache.commons.math.stat.descriptive.StatisticalSummary, org.apache.commons.math.stat.descriptive.StatisticalSummary)
173 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticTTest(double[], double[], double)
182 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticTTest(double[], double[])
190 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticTTest(org.apache.commons.math.stat.descriptive.StatisticalSummary, org.apache.commons.math.stat.descriptive.StatisticalSummary)
199 * @see org.apache.commons.math.stat.inference.TTest#pairedT(double[], double[])
207 * @see org.apache.commons.math.stat.inference.TTest#pairedTTest(double[], double[], double
    [all...]
ChiSquareTestImpl.java 17 package org.apache.commons.math.stat.inference;
34 /** Distribution used to compute inference statistics. */
46 * inference statistics.
47 * @param x distribution used to compute inference statistics.
415 * Modify the distribution used to compute inference statistics.
TTestImpl.java 17 package org.apache.commons.math.stat.inference;
38 /** Distribution used to compute inference statistics.
53 * inference statistics.
54 * @param t distribution used to compute inference statistics.
    [all...]
  /packages/apps/Dialer/java/com/android/dialer/p13n/inference/
P13nRanking.java 17 package com.android.dialer.p13n.inference;
26 import com.android.dialer.p13n.inference.protocol.P13nRanker;
27 import com.android.dialer.p13n.inference.protocol.P13nRankerFactory;
  /external/desugar/test/java/com/google/devtools/build/android/desugar/
ByteCodeTypePrinter.java 88 BytecodeTypeInference inference = new BytecodeTypeInference(access, internalName, name, desc); local
89 mv = new MethodIrTypeDumper(mv, inference, printWriter);
90 inference.setDelegateMethodVisitor(mv);
91 // Let the type inference runs first.
92 return inference;
109 private final BytecodeTypeInference inference; field in class:ByteCodeTypePrinter.MethodIrTypeDumper
114 MethodVisitor visitor, BytecodeTypeInference inference, PrintWriter printWriter) {
116 this.inference = inference;
121 printer.print(" |__STACK: " + inference.getOperandStackAsString() + "\n")
    [all...]
  /external/desugar/java/com/google/devtools/build/android/desugar/
TryWithResourcesRewriter.java 203 BytecodeTypeInference inference = null; local
210 inference = new BytecodeTypeInference(access, internalName, name, desc);
211 inference.setDelegateMethodVisitor(visitor);
212 visitor = inference;
216 new TryWithResourceVisitor(internalName, name + desc, visitor, classLoader, inference);
288 + "but the type inference is null.",
  /external/tensorflow/tensorflow/examples/tutorials/mnist/
mnist.py 18 Implements the inference/loss/training pattern for model building.
20 1. inference() - Builds the model as far as required for running the network
22 2. loss() - Adds to the inference model the layers required to generate loss.
45 def inference(images, hidden1_units, hidden2_units): function
46 """Build the MNIST model up to where it may be used for inference.
fully_connected_feed.py 128 # Build a Graph that computes predictions from the inference model.
129 logits = mnist.inference(images_placeholder,
  /external/icu/icu4c/source/samples/ufortune/resources/
fortune_resources.mak 37 # File name extensions for inference rule matching.
44 # Inference rule, for compiling a .txt file into a .res file.
  /external/tensorflow/tensorflow/contrib/factorization/examples/
mnist.py 121 def inference(inp, num_clusters, hidden1_units, hidden2_units): function
122 """Build the MNIST model up to where it may be used for inference.
195 # Build a Graph that computes predictions from the inference model.
196 logits, clustering_loss, kmeans_init, kmeans_training_op = inference(
  /external/tensorflow/tensorflow/contrib/model_pruning/examples/cifar10/
cifar10_eval.py 112 # inference model.
113 logits = cifar10.inference(images)
cifar10_train.py 64 # inference model.
65 logits = cifar10.inference(images)
cifar10_pruning.py 23 # Compute inference on the model inputs to make a prediction.
24 predictions = inference(inputs)
170 def inference(images): function
278 logits: Logits from inference().
  /external/tensorflow/tensorflow/examples/how_tos/reading_data/
fully_connected_preloaded.py 66 # Build a Graph that computes predictions from the inference model.
67 logits = mnist.inference(images, FLAGS.hidden1, FLAGS.hidden2)
fully_connected_preloaded_var.py 72 # Build a Graph that computes predictions from the inference model.
73 logits = mnist.inference(images, FLAGS.hidden1, FLAGS.hidden2)
fully_connected_reader.py 136 # Build a Graph that computes predictions from the inference model.
137 logits = mnist.inference(image_batch, FLAGS.hidden1, FLAGS.hidden2)
  /external/protobuf/js/
debug.js 85 var message = thing; // Copy because we don't want type inference on thing.
  /external/tensorflow/tensorflow/compiler/xla/tools/parser/
hlo_parser_test.cc 552 // batch norm inference
563 ROOT %batch-norm-inference = f32[2,2,2,2]{3,2,1,0} batch-norm-inference(f32[2,2,2,2]{3,2,1,0} %input, f32[2]{0} %offset, f32[2]{0} %scale, f32[2]{0} %mean, f32[2]{0} %variance), epsilon=0.001, feature_index=0
    [all...]

Completed in 343 milliseconds

1 2