/device/linaro/bootloader/edk2/MdeModulePkg/Bus/Usb/UsbMassStorageDxe/ |
UsbMassStorageDxe.inf | 4 # The USB mass storage class is specified in two layers: the bottom layer
5 # is the transportation protocol. The top layer is the command set.
6 # The transportation layer provides the transportation of the command, data and result.
|
/external/iproute2/man/man8/ |
ip-vrf.8 | 32 A VRF provides traffic isolation at layer 3 for routing, similar to how a 33 VLAN is used to isolate traffic at layer 2. Fundamentally, a VRF is a separate 81 The VRF helper *only* affects network layer sockets.
|
/external/libavc/encoder/ |
ih264e_cavlc.h | 56 * The mb syntax layer for intra slices constitutes luma mb mode, luma sub modes 79 * The mb syntax layer for inter slices constitutes luma mb mode, luma sub modes 102 * The mb syntax layer for inter slices constitutes luma mb mode, luma sub modes
|
/external/libvpx/libvpx/vpx/ |
svc_context.h | 65 int width; // width of highest layer 66 int height; // height of highest layer 71 int layer; member in struct:SvcInternal
|
/external/perfetto/include/perfetto/tracing/core/ |
consumer.h | 34 // Called by Service (or more typically by the transport layer, on behalf of 39 // Called by the Service or by the transport layer if the connection with the 53 // Called back by the Service (or transport layer) after invoking
|
/external/swiftshader/src/OpenGL/libGLESv2/ |
Framebuffer.h | 42 void setColorbuffer(GLenum type, GLuint colorbuffer, GLuint index, GLint level = 0, GLint layer = 0); 43 void setDepthbuffer(GLenum type, GLuint depthbuffer, GLint level = 0, GLint layer = 0); 44 void setStencilbuffer(GLenum type, GLuint stencilbuffer, GLint level = 0, GLint layer = 0);
|
/external/tensorflow/tensorflow/contrib/crf/ |
README.md | 3 The CRF module implements a linear-chain CRF layer for learning to predict tag sequences. This variant of the CRF is factored into unary potentials for every element in the sequence and binary potentials for every transition between output tags. 7 Below is an example of the API, which learns a CRF for some random data. The linear layer in the example can be replaced by any neural network. 37 # Compute unary scores from a linear layer.
|
/external/tensorflow/tensorflow/contrib/specs/ |
README.md | 9 - operator `|` takes the output from one layer and "pipes" it into the next 10 - operator `**` repeats a layer multiple times 24 is a conjunction of a base layer and the activation. For example, `Fs` 25 represents a fully connected layer followed by a sigmoid, whereas `Ft` 26 represents a fully connected layer followed by a Tanh. 34 Nonlinearities (suffixes for C/F, so Cs = convolutional layer + sigmoid): 109 layer types; it's not a replacement for the standard TensorFlow or 110 slim APIs. If you have some complex layer type or construct that's 221 - Seq(a, b, c).add(name=layer).add(name=layer) for explicit seq. structure [all...] |
/frameworks/base/libs/hwui/renderthread/ |
DrawFrameTask.h | 55 * and contains state (such as layer updaters & new DisplayLists) that is 69 void pushLayerUpdate(DeferredLayerUpdater* layer); 70 void removeLayerUpdate(DeferredLayerUpdater* layer);
|
/hardware/intel/img/hwcomposer/moorefield_hdmi/common/base/ |
DisplayAnalyzer.cpp | 118 hwc_layer_1_t *layer = &display->hwLayers[i]; local 119 if (layer) layer->compositionType = type;
|
/hardware/interfaces/graphics/composer/2.1/utils/hwc2onfbadapter/include/hwc2onfbadapter/ |
HWC2OnFbAdapter.h | 67 bool removeLayer(hwc2_layer_t layer); 68 bool hasLayer(hwc2_layer_t layer) const; 69 bool markLayerDirty(hwc2_layer_t layer, bool dirty);
|
/hardware/libhardware_legacy/include/hardware_legacy/ |
tdls.h | 14 WIFI_TDLS_FAILED /* TDLS permanent failed. Inform error to upper layer 61 * If upper layer has no specific values, this should be NULL 63 * handler is used to inform the upper layer about the status change and the corresponding reason
|
/hardware/qcom/camera/msm8998/QCamera2/stack/common/ |
mm_camera_shim.h | 36 * MCT shim layer APIs 59 * Command types to process in shim layer 75 * Structure to communicate command with shim layer
|
/packages/apps/UnifiedEmail/src/com/android/mail/browse/ |
InlineAttachmentViewIntentBuilderCreatorHolder.java | 24 * Unfortunately, this pattern requires three layers. The holder (the top layer) is created at 30 * The middle layer is the {@link InlineAttachmentViewIntentBuilderCreator}. Only one of 37 * The bottom layer is the {@link InlineAttachmentViewIntentBuilder}. This builder
|
/packages/services/Car/car-lib/src/android/car/vms/ |
VmsAssociatedLayer.java | 25 * A VMS Layer with a list of publisher IDs it is associated with. 37 public VmsAssociatedLayer(VmsLayer layer, Set<Integer> publisherIds) { 38 mLayer = layer;
|
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.15-4.8/sysroot/usr/include/linux/ |
net_tstamp.h | 105 /* PTP v2/802.AS1, any layer, any kind of event packet */ 107 /* PTP v2/802.AS1, any layer, Sync packet */ 109 /* PTP v2/802.AS1, any layer, Delay_req packet */
|
/system/bt/hci/include/ |
hci_layer.h | 71 // Set the callback that the HCI layer uses to send data upwards 76 // Send a command through the HCI layer 83 // Send some data downward through the HCI layer
|
/external/tensorflow/tensorflow/contrib/layers/python/layers/ |
layers.py | 116 layer = pooling_layers.AveragePooling2D( 122 outputs = layer.apply(inputs) 165 layer = pooling_layers.AveragePooling3D( 171 outputs = layer.apply(inputs) 192 """Adds a Batch Normalization layer from http://arxiv.org/abs/1502.03167. 227 not used. When the next layer is linear (also e.g. `nn.relu`), this can be 228 disabled since the scaling can be done by the next layer. 239 is_training: Whether or not the layer is in training mode. In training mode 244 reuse: Whether or not the layer and its variables should be reused. To be 245 able to reuse the layer scope must be given [all...] |
/hardware/qcom/display/msm8226/libhwcomposer/ |
hwc_utils.cpp | 685 bool isDownscaleRequired(hwc_layer_1_t const* layer) { 686 hwc_rect_t displayFrame = layer->displayFrame; 687 hwc_rect_t sourceCrop = integerizeSourceCrop(layer->sourceCropf); 699 bool needsScaling(hwc_layer_1_t const* layer) { 701 hwc_rect_t displayFrame = layer->displayFrame; 702 hwc_rect_t sourceCrop = integerizeSourceCrop(layer->sourceCropf); 715 // Checks if layer needs scaling with split 716 bool needsScalingWithSplit(hwc_context_t* ctx, hwc_layer_1_t const* layer, 727 hwc_rect_t sourceCrop = integerizeSourceCrop(layer->sourceCropf); 728 hwc_rect_t displayFrame = layer->displayFrame 806 hwc_layer_1_t *layer = &list->hwLayers[i]; local 842 hwc_layer_1_t const* layer = &list->hwLayers[i]; local 1198 hwc_layer_1_t* layer = (hwc_layer_1_t*)&list->hwLayers[j]; local 2083 hwc_layer_1_t* layer = &(list->hwLayers[index]); local [all...] |
hwc_utils.h | 119 int numAppLayers; //Total - 1, excluding FB layer. 132 bool secureUI; // Secure display layer 158 uint32_t mFlags; //qcom specific layer flags 199 void add(hwc_layer_1_t* layer, overlay::Rotator *rot); 200 //Resets the mapping of layer to rotator 260 bool isSecuring(hwc_context_t* ctx, hwc_layer_1_t const* layer); 262 // Returns true, if the input layer format is supported by rotator 264 //Returns true, if the layer is YUV or the layer has been rendered by CPU 267 bool isAlphaScaled(hwc_layer_1_t const* layer); [all...] |
/hardware/qcom/display/msm8996/sdm/libs/core/ |
display_base.cpp | 150 std::vector<Layer *>layers = layer_stack->layers; 162 DLOGE("Either layer count is zero or GPU target layer is not present"); 168 // Check GPU target layer 169 Layer *gpu_target_layer = layers.at(gpu_target_index); 172 DLOGE("Invalid src rect for GPU target layer"); 177 DLOGE("Invalid dst rect for GPU target layer"); 196 DLOGE("GPU target layer dst rect is not with in limits gpu wxh %fx%f mixer wxh %dx%d", 296 // Layer stack attributes has changed, need to Reconfigure, currently in use for Hybrid Comp 580 Layer *layer = hw_layers_.info.stack->layers.at(layer_index) local 959 Layer *layer = layers.at(i); local 977 Layer *layer = layers.at(max_area_layer_index); local [all...] |
/system/bt/stack/l2cap/ |
l2c_csm.cc | 170 /* Disconnect unless ACL collision and upper layer wants to handle it */ 275 case L2CEVT_L2CA_DATA_WRITE: /* Upper layer data to send */ 369 case L2CEVT_L2CA_DATA_WRITE: /* Upper layer data to send */ 453 case L2CEVT_L2CA_DATA_WRITE: /* Upper layer data to send */ 588 case L2CEVT_L2CA_DATA_WRITE: /* Upper layer data to send */ 686 case L2CEVT_L2CA_DATA_WRITE: /* Upper layer data to send */ 699 /* We have feature info, so now give the upper layer connect IND */ [all...] |
/external/tensorflow/tensorflow/python/layers/ |
convolutional.py | 17 """Contains the convolutional layer classes and their functional aliases. 35 class _Conv(base.Layer): 36 """Abstract nD convolution layer (private, used as implementation base). 38 This layer creates a convolution kernel that is convolved 39 (actually cross-correlated) with the layer input to produce a tensor of 66 use_bias: Boolean, whether the layer uses a bias. 75 norm constraints or value constraints for layer weights). The function 83 name: A string, the name of the layer. 228 """1D convolution layer (e.g. temporal convolution). 230 This layer creates a convolution kernel that is convolve [all...] |
/hardware/qcom/display/msm8084/libhwcomposer/ |
hwc_utils.cpp | 698 bool isDownscaleRequired(hwc_layer_1_t const* layer) { 699 hwc_rect_t displayFrame = layer->displayFrame; 700 hwc_rect_t sourceCrop = integerizeSourceCrop(layer->sourceCropf); 712 bool needsScaling(hwc_layer_1_t const* layer) { 714 hwc_rect_t displayFrame = layer->displayFrame; 715 hwc_rect_t sourceCrop = integerizeSourceCrop(layer->sourceCropf); 728 // Checks if layer needs scaling with split 729 bool needsScalingWithSplit(hwc_context_t* ctx, hwc_layer_1_t const* layer, 740 hwc_rect_t sourceCrop = integerizeSourceCrop(layer->sourceCropf); 741 hwc_rect_t displayFrame = layer->displayFrame 819 hwc_layer_1_t *layer = &list->hwLayers[i]; local 882 hwc_layer_1_t const* layer = &list->hwLayers[i]; local 1210 hwc_layer_1_t* layer = (hwc_layer_1_t*)&list->hwLayers[j]; local 2083 hwc_layer_1_t* layer = &(list->hwLayers[index]); local [all...] |
/external/libpcap/pcap/ |
dlt.h | 45 * Link-layer header type codes. 59 * for detailed descriptions of some of these link-layer header types. 115 * as 117 so that pflog captures would use a link-layer header type 161 * a link-layer type of 99 for the tcpdump it supplies. The link-layer 170 * link-layer header type LINKTYPE_ values corresponding to DLT_ types 176 * Values starting with 104 are used for newly-assigned link-layer 177 * header type values; for those link-layer header types, the DLT_ 210 * so that we don't have to worry about the link-layer header.) 223 * that the AF_ type in the link-layer header is in network byte order [all...] |