HomeSort by relevance Sort by last modified time
    Searched refs:dataset (Results 26 - 50 of 208) sorted by null

12 3 4 5 6 7 8 9

  /external/protobuf/benchmarks/
generate_datasets.cc 43 const char *file_prefix = "dataset.";
60 std::cerr << "For dataset " << name << ", no such message: "
69 std::cerr << "For dataset " << name << ", payload[" << i << "] fails "
75 BenchmarkDataset dataset; local
76 dataset.set_name(name);
77 dataset.set_message_name(message_name);
79 dataset.add_payload()->assign(payload[i]);
85 dataset.SerializeToOstream(&writer);
88 std::cerr << "Wrote dataset: " << fname << "\n";
  /external/tensorflow/tensorflow/contrib/data/python/ops/
interleave_ops.py 15 """Non-deterministic dataset transformations."""
31 class ParallelInterleaveDataset(dataset_ops.Dataset):
32 """A `Dataset` that maps a function over its input and flattens the result."""
56 dataset = map_func(*nested_args)
58 dataset = map_func(nested_args)
60 if not isinstance(dataset, dataset_ops.Dataset):
61 raise TypeError("`map_func` must return a `Dataset` object.")
63 self._output_classes = dataset.output_classes
64 self._output_types = dataset.output_type
    [all...]
dataset_ops.py 31 class Dataset(dataset_ops.Dataset):
34 A `Dataset` can be used to represent an input pipeline as a
39 def __init__(self, dataset):
40 super(Dataset, self).__init__()
41 self._dataset = dataset
63 @deprecation.deprecated(None, "Use `tf.data.Dataset.from_tensors()`.")
65 """Creates a `Dataset` with a single element, comprising the given tensors.
71 A `Dataset`.
73 return Dataset(dataset_ops.TensorDataset(tensors)
    [all...]
stats_ops.py 33 in this module when defining your @{tf.data.Dataset}. All statistics will be
36 produced by iterating over a dataset:
39 dataset = ...
40 dataset = dataset.apply(stats_ops.bytes_produced_stats("total_bytes"))
47 dataset = ...
48 iterator = dataset.make_one_shot_iterator()
110 """Records the number of bytes produced by each element of the input dataset.
113 over the output dataset.
120 A `Dataset` transformation function, which can be passed t
    [all...]
error_ops.py 15 """Ignore_errors dataset transformations."""
28 """Creates a `Dataset` from another `Dataset` and silently ignores any errors.
30 Use this transformation to produce a dataset that contains the same elements
35 dataset = tf.data.Dataset.from_tensor_slices([1., 2., 0., 4.])
38 dataset = dataset.map(lambda x: tf.check_numerics(1. / x, "error"))
41 dataset =
42 dataset.apply(tf.contrib.data.ignore_errors()) # ==> { 1., 0.5, 0.2
    [all...]
unique.py 15 """Unique element dataset transformations."""
28 """Creates a `Dataset` from another `Dataset`, discarding duplicates.
30 Use this transformation to produce a dataset that contains one instance of
34 dataset = tf.data.Dataset.from_tensor_slices([1, 37, 2, 37, 2, 1])
37 dataset = dataset.apply(tf.contrib.data.unique()) # ==> { 1, 37, 2 }
41 A `Dataset` transformation function, which can be passed to
42 @{tf.data.Dataset.apply}
    [all...]
  /external/tensorflow/tensorflow/contrib/eager/python/examples/linear_regression/
linear_regression_test.py 53 dataset = linear_regression.synthetic_dataset(true_w, true_b, noise_level,
56 it = tfe.Iterator(dataset)
71 dataset = linear_regression.synthetic_dataset(
76 linear_regression.fit(model, dataset, optimizer, logdir=self._tmp_logdir)
89 dataset = linear_regression.synthetic_dataset(
95 burn_in_dataset = dataset.take(10)
107 linear_regression.fit(model, dataset, optimizer)
  /external/tensorflow/tensorflow/python/data/ops/
iterator_ops.py 54 """Represents the state of iterating through a `Dataset`."""
61 instead use `Dataset.make_initializable_iterator()` or
62 `Dataset.make_one_shot_iterator()`.
70 each component of an element of this dataset.
72 corresponding to each component of an element of this dataset.
96 The returned iterator is not bound to a particular dataset, and it has
98 `Iterator.make_initializer(dataset)`.
105 dataset_range = Dataset.range(10)
138 each component of an element of this dataset.
140 corresponding to each component of an element of this dataset. I
    [all...]
  /external/tensorflow/tensorflow/contrib/data/python/kernel_tests/
stats_dataset_ops_test.py 53 dataset = dataset_ops.Dataset.range(100).map(
56 iterator = dataset.make_initializable_iterator()
79 dataset = dataset_ops.Dataset.range(100).apply(
81 iterator = dataset.make_initializable_iterator()
98 dataset = dataset_ops.Dataset.range(100).apply(
100 iterator = dataset.make_initializable_iterator()
120 dataset = dataset_ops.Dataset.range(100).apply
    [all...]
scan_dataset_op_test.py 37 return dataset_ops.Dataset.from_tensors(0).repeat(None).apply(
61 iterator = dataset_ops.Dataset.from_tensors(1).repeat(None).apply(
85 dataset = dataset_ops.Dataset.from_tensors(0).repeat(5).apply(
87 self.assertEqual([None], dataset.output_shapes[0][0].as_list())
88 self.assertIs(None, dataset.output_shapes[0][1].ndims)
89 self.assertEqual([], dataset.output_shapes[1].as_list())
91 iterator = dataset.make_one_shot_iterator()
107 dataset = dataset_ops.Dataset.range(10
    [all...]
  /external/tensorflow/tensorflow/core/framework/
dataset.h 71 // Wrapper around GraphDefBuilder. Used to serialize Dataset graph.
122 Status AddDataset(const GraphDatasetBase* dataset,
124 return AddDataset(dataset, inputs, {}, output);
135 Status AddDataset(const GraphDatasetBase* dataset,
143 return AddDataset(dataset, enumerated_inputs, {}, attrs, output);
147 const GraphDatasetBase* dataset,
181 // TODO(b/65524810): Hack to allow functions to capture Dataset op
199 // Uses a heuristic to whitelist source dataset ops which have been
204 return (StringPiece(op_def->name()).ends_with("Dataset") &&
207 dataset::WhitelistedStatefulOpRegistry::Global()->Contains
488 const DatasetType* dataset; member in struct:tensorflow::DatasetIterator::Params
501 const DatasetType* dataset() const { return params_.dataset; } function in class:tensorflow::DatasetIterator
    [all...]
dataset.cc 15 #include "tensorflow/core/framework/dataset.h"
37 // Transfers ownership of `dataset` to `*this`.
38 explicit DatasetVariantWrapper(DatasetBase* dataset) : dataset_(dataset) {}
76 const GraphDatasetBase* dataset,
81 const string& op_type_name = dataset->op_name();
90 opts->WithAttr("output_shapes", dataset->output_shapes())));
94 opts->WithAttr("output_types", dataset->output_dtypes())));
217 "Dataset tensor must be a scalar of dtype DT_VARIANT.");
222 return errors::InvalidArgument("Tensor must be a Dataset object.")
242 DatasetBase* dataset = nullptr; local
    [all...]
  /external/tensorflow/tensorflow/contrib/tensor_forest/kernels/v4/
decision_node_evaluator.cc 61 const std::unique_ptr<TensorDataSet>& dataset, int example) const {
62 const float val = dataset->GetExampleValue(example, feature_num_);
83 const std::unique_ptr<TensorDataSet>& dataset, int example) const {
87 dataset->GetExampleValue(example, feature_num_[i]);
108 const std::unique_ptr<TensorDataSet>& dataset, int example) const {
109 const float val = dataset->GetExampleValue(example, feature_num_);
  /external/tensorflow/tensorflow/core/kernels/data/
range_dataset_op.cc 17 #include "tensorflow/core/kernels/data/dataset.h"
42 *output = new Dataset(ctx, start, stop, step);
46 class Dataset : public GraphDatasetBase {
48 Dataset(OpKernelContext* ctx, int64 start, int64 stop, int64 step)
70 step_, ")::Dataset");
87 class Iterator : public DatasetIterator<Dataset> {
90 : DatasetIterator<Dataset>(params) {
91 next_ = params.dataset->start_;
98 if ((dataset()->step_ > 0 && next_ >= dataset()->stop_) |
    [all...]
dataset_utils.cc 20 namespace dataset { namespace in namespace:tensorflow
37 // Retrieve the dataset that was created in `f`.
42 // Create an iterator for the dataset that was returned by `f`.
48 } // namespace dataset
map_and_batch_dataset_op.cc 21 #include "tensorflow/core/kernels/data/dataset.h"
73 *output = new Dataset(input, batch_size, num_parallel_batches,
79 class Dataset : public DatasetBase {
81 Dataset(const DatasetBase* input, int64 batch_size,
96 ~Dataset() override { input_->Unref(); }
112 string DebugString() override { return "MapAndBatchDatasetOp::Dataset"; }
115 class Iterator : public DatasetIterator<Dataset> {
118 : DatasetIterator<Dataset>(params),
119 input_impl_(params.dataset->input_->MakeIterator(params.prefix)),
120 invocation_results_(params.dataset->batch_size_
    [all...]
reader_dataset_ops.cc 17 #include "tensorflow/core/kernels/data/dataset.h"
76 *output = new Dataset(ctx, std::move(filenames), compression_type,
81 class Dataset : public GraphDatasetBase {
83 Dataset(OpKernelContext* ctx, std::vector<string> filenames,
109 string DebugString() override { return "TextLineDatasetOp::Dataset"; }
127 class Iterator : public DatasetIterator<Dataset> {
130 : DatasetIterator<Dataset>(params) {}
160 if (current_file_index_ == dataset()->filenames_.size()) {
209 if (current_file_index_ >= dataset()->filenames_.size()) {
212 " >= filenames_.size():", dataset()->filenames_.size())
    [all...]
  /developers/samples/android/input/autofill/AutofillFramework/afservice/src/main/java/com/example/android/autofill/service/data/
FakeAutofillDataBuilder.java 47 "dataset-" + datasetNumber + "." + partition, mPackageName);
60 AutofillDataset dataset, int partition) {
63 datasetWithFilledAutofillFields.autofillDataset = dataset;
69 mSeed, dataset.getId());
  /packages/experimental/FillService/src/foo/bar/fill/
AuthActivity.java 9 import android.service.autofill.Dataset;
31 final Dataset dataset; local
41 .addDataset(new Dataset.Builder(presentation1)
47 .addDataset(new Dataset.Builder(presentation2)
54 dataset = null;
59 dataset = new Dataset.Builder(presentation)
74 result.putExtra(AutofillManager.EXTRA_AUTHENTICATION_RESULT, dataset);
  /cts/tests/autofillservice/src/android/autofillservice/cts/
ManualAuthenticationActivity.java 38 public static void setDataset(CannedFillResponse.CannedDataset dataset) {
39 sDataset = dataset;
61 throw new IllegalStateException("no dataset or response");
  /developers/build/prebuilts/gradle/SpeedTracker/Wearable/src/main/java/com/example/android/wearable/speedtracker/ui/
SpeedPickerListAdapter.java 38 public SpeedPickerListAdapter(Context context, int[] dataset) {
41 mDataSet = dataset;
  /developers/samples/android/wearable/wear/SpeedTracker/Wearable/src/main/java/com/example/android/wearable/speedtracker/ui/
SpeedPickerListAdapter.java 38 public SpeedPickerListAdapter(Context context, int[] dataset) {
41 mDataSet = dataset;
  /development/samples/browseable/SpeedTracker/Wearable/src/com.example.android.wearable.speedtracker/ui/
SpeedPickerListAdapter.java 38 public SpeedPickerListAdapter(Context context, int[] dataset) {
41 mDataSet = dataset;
  /external/tensorflow/tensorflow/contrib/eager/python/examples/mnist/
mnist_graph_test.py 43 # Create a model, optimizer, and dataset as would be done
47 dataset = tf.data.Dataset.from_tensors((images, labels))
51 (images, labels) = dataset.make_one_shot_iterator().get_next()
  /external/tensorflow/tensorflow/contrib/kfac/examples/
mnist.py 34 """Loads MNIST dataset into memory.
38 num_epochs: int. Number of passes to make over the dataset.
42 use_fake_data: bool. If True, generate a synthetic dataset rather than
66 dataset = tf.data.Dataset.from_tensor_slices((np.asarray(
68 return (dataset.repeat(num_epochs).shuffle(num_examples).batch(batch_size)

Completed in 647 milliseconds

12 3 4 5 6 7 8 9