/prebuilts/clang/host/linux-x86/clang-4393122/include/llvm/ExecutionEngine/Orc/ |
ObjectTransformLayer.h | 22 /// @brief Object mutating layer. 24 /// This layer accepts sets of ObjectFiles (via addObjectSet). It 26 /// the set of transformed objects to the layer below. 39 /// add the resulting set of objects to the base layer, along with the 68 /// base layer's implementation.
|
/prebuilts/clang/host/linux-x86/clang-4479392/include/llvm/ExecutionEngine/Orc/ |
IRTransformLayer.h | 22 /// @brief IR mutating layer. 24 /// This layer accepts sets of LLVM IR Modules (via addModuleSet). It 26 /// the set of transformed modules to the layer below. 39 /// add the resulting set of modules to the base layer, along with the 69 /// base layer's implementation.
|
ObjectTransformLayer.h | 22 /// @brief Object mutating layer. 24 /// This layer accepts sets of ObjectFiles (via addObjectSet). It 26 /// the set of transformed objects to the layer below. 39 /// add the resulting set of objects to the base layer, along with the 68 /// base layer's implementation.
|
/prebuilts/clang/host/linux-x86/clang-4579689/include/llvm/ExecutionEngine/Orc/ |
ObjectTransformLayer.h | 25 /// @brief Object mutating layer. 27 /// This layer accepts sets of ObjectFiles (via addObject). It 29 /// the set of transformed objects to the layer below. 42 /// add the resulting set of objects to the base layer, along with the 65 /// base layer's implementation.
|
/prebuilts/clang/host/linux-x86/clang-4630689/include/llvm/ExecutionEngine/Orc/ |
ObjectTransformLayer.h | 25 /// @brief Object mutating layer. 27 /// This layer accepts sets of ObjectFiles (via addObject). It 29 /// the set of transformed objects to the layer below. 42 /// add the resulting set of objects to the base layer, along with the 65 /// base layer's implementation.
|
/prebuilts/clang/host/linux-x86/clang-4639204/include/llvm/ExecutionEngine/Orc/ |
ObjectTransformLayer.h | 25 /// @brief Object mutating layer. 27 /// This layer accepts sets of ObjectFiles (via addObject). It 29 /// the set of transformed objects to the layer below. 42 /// add the resulting set of objects to the base layer, along with the 65 /// base layer's implementation.
|
/prebuilts/clang/host/linux-x86/clang-4691093/include/llvm/ExecutionEngine/Orc/ |
ObjectTransformLayer.h | 25 /// @brief Object mutating layer. 27 /// This layer accepts sets of ObjectFiles (via addObject). It 29 /// the set of transformed objects to the layer below. 42 /// add the resulting set of objects to the base layer, along with the 65 /// base layer's implementation.
|
/external/deqp/external/vulkancts/modules/vulkan/geometry/ |
vktGeometryLayeredRenderingTests.cpp | 59 TEST_TYPE_DEFAULT_LAYER, // !< draw to default layer 60 TEST_TYPE_SINGLE_LAYER, // !< draw to single layer 64 TEST_TYPE_INVOCATION_PER_LAYER, // !< draw to all layers, one invocation per layer 408 inline tcu::ConstPixelBufferAccess getLayer (const int layer) const 410 return tcu::getSubregion(m_wholeImage, 0, (m_1dModifier * layer), ((~m_1dModifier & 1) * layer), m_width, m_height, 1); 499 << tcu::TestLog::ImageSet("LayerContent", "Layer content") 500 << tcu::TestLog::Image("Layer", "Layer", image) 507 << tcu::TestLog::ImageSet("LayerContent", "Layer content" 603 str << "slice z = " << layer; local 605 str << "layer " << layer; local [all...] |
/external/tensorflow/tensorflow/docs_src/get_started/ |
custom_estimators.md | 81 and a logits output layer. 175 * An [input layer](https://developers.google.com/machine-learning/glossary/#input_layer) 177 * An [output layer](https://developers.google.com/machine-learning/glossary/#output_layer) 179 ### Define the input layer 191 creating the model's input layer. 195 alt="A diagram of the input layer, in this case a 1:1 mapping from raw-inputs to features." 206 dimensions defined by `params['hidden_layers']`. In a `dense` layer each node 207 is connected to every node in the preceding layer. Here's the relevant code: 215 * The `units` parameter defines the number of output neurons in a given layer. 220 The variable `net` here signifies the current top layer of the network. Durin [all...] |
/external/webrtc/webrtc/modules/video_coding/codecs/test/ |
videoprocessor_integrationtest.cc | 116 // Some quantities defined per temporal layer (at most 3 layers in this test). 238 // Update layer per-frame-bandwidth. 248 // layer frame in this test. 263 // Update layer data. 278 // Encoding bitrate per layer: from the start of the update/run to the 329 printf("Rates statistics for Layer data \n"); 331 printf("Layer #%d \n", i); 338 " Target Layer Bit rate: %f \n" 339 " Layer frame rate: %f, \n" 340 " Layer per frame bandwidth: %f, \n [all...] |
/external/tensorflow/tensorflow/contrib/kfac/python/ops/ |
fisher_blocks.py | 202 params: The parameters of this layer (Tensor or tuple of Tensors). 266 params: The parameters of this layer (Tensor or tuple of Tensors). 319 connected layer. Unlike NaiveDiagonalFB this uses the low-variance "sum of 328 Consider fully connected layer in this model with (unshared) weight matrix 329 'w'. For an example 'x' that produces layer inputs 'a' and output 335 to the layer's parameters 'w'. 367 [input_size, output_size] corresponding to layer's weights. If not, a 369 to the layer's bias. 385 [input_size, output_size] corresponding to layer's weights. If not, a 387 to the layer's bias [all...] |
/external/scapy/scapy/ |
config.py | 109 def register(self, num, layer): 110 self.register_num2layer(num, layer) 111 self.register_layer2num(num, layer) 113 def register_num2layer(self, num, layer): 114 self.num2layer[num] = layer 115 def register_layer2num(self, num, layer): 116 self.layer2num[layer] = num 133 for num,layer in six.iteritems(self.num2layer): 134 if layer in self.layer2num and self.layer2num[layer] == num [all...] |
/frameworks/base/graphics/java/android/graphics/drawable/ |
RippleDrawable.java | 55 * special mask layer that is not drawn to the screen. A single layer may be 57 * {@link android.R.id#mask}. At run time, a single layer may be set as the 58 * mask using {@code setId(..., android.R.id.mask)} or an existing mask layer 68 * If a mask layer is set, the ripple effect will be masked against that layer 71 * If no mask layer is set, the ripple effect is masked against the composite 125 /** The masking layer, e.g. the layer with id R.id.mask. */ 344 // If the layer is bounded, then we don't need to project [all...] |
/device/linaro/bootloader/edk2/NetworkPkg/UefiPxeBcDxe/ |
PxeBcBoot.h | 55 @param[in] Layer Pointer to option boot item layer.
71 IN UINT16 *Layer,
|
/external/kernel-headers/original/uapi/linux/ |
net_tstamp.h | 126 /* PTP v2/802.AS1, any layer, any kind of event packet */ 128 /* PTP v2/802.AS1, any layer, Sync packet */ 130 /* PTP v2/802.AS1, any layer, Delay_req packet */
|
/external/libnl/lib/route/cls/ |
ematch_grammar.l | 99 [eE][tT][hH] { yylval->i = TCF_LAYER_LINK; return LAYER; } 102 [iI][pP] { yylval->i = TCF_LAYER_NETWORK; return LAYER; } 104 [tT][cC][pP] { yylval->i = TCF_LAYER_TRANSPORT; return LAYER; }
|
/external/libpcap/ |
pcap_dump_open.3pcap | 64 The time stamp precision, link-layer type, and snapshot length from 66 are used as the link-layer type and snapshot length of the output file. 73 opening the file, and has the same time stamp precision, link-layer
|
pcap_dump_open.3pcap.in | 64 The time stamp precision, link-layer type, and snapshot length from 66 are used as the link-layer type and snapshot length of the output file. 73 opening the file, and has the same time stamp precision, link-layer
|
/external/mdnsresponder/mDNSPosix/ |
ExampleClientApp.c | 28 #include "mDNSEmbeddedAPI.h" // Defines the interface to the client layer above 67 // 3. Give the mDNSPosix layer a chance to add its information to the fd_set and timeout 81 // 5. Call mDNSPosixProcessFDSet to let the mDNSPosix layer do its work
|
/external/mesa3d/src/gallium/state_trackers/nine/ |
surface9.h | 46 unsigned layer; member in struct:NineSurface9 70 unsigned Layer, 82 unsigned Layer,
|
/external/scapy/scapy/contrib/ |
coap.uts | 1 % CoAP layer test campaign 4 = Import the CoAP layer 38 + Test layer binding
|
nsh.uts | 2 * Tests for the Scapy NSH layer 5 = Import the nsh layer 8 + Basic Layer Tests
|
/external/swiftshader/src/OpenGL/libGLESv2/ |
Device.hpp | 66 void setRenderTarget(int index, egl::Image *renderTarget, unsigned int layer); 67 void setDepthBuffer(egl::Image *depthBuffer, unsigned int layer); 68 void setStencilBuffer(egl::Image *stencilBuffer, unsigned int layer);
|
/external/tensorflow/tensorflow/contrib/kfac/ |
README.md | 22 are the weights for layer i?"). As such, you must add some additional code while 31 in **faster training**. For an 8-layer Autoencoder, K-FAC converges to the same 52 1. Registering layer inputs, weights, and pre-activations with a
|
/external/tensorflow/tensorflow/contrib/receptive_field/python/util/ |
parse_layer_parameters.py | 25 # White-listed layer operations, which do not affect the receptive field 82 ValueError: If the weight layer node is invalid. 87 "Weight layer's name input to conv layer does not end with '/read'") 91 raise ValueError("Weight layer is not of type VariableV2") 214 ValueError: If padding layer is invalid. 218 raise ValueError("Padding layer name does not end with '/paddings'") 236 """Gets layer parameters relevant for RF computation. 251 input feature map to this layer. If set to None, then the padding may be 254 force: If True, the function does not raise a ValueError if the layer op i [all...] |