/external/webrtc/webrtc/video/ |
screenshare_loopback.cc | 133 DEFINE_int32(num_spatial_layers, 1, "Number of spatial layers to use."); 140 "Spatial layer to show or analyze. -1 to disable filtering.");
|
video_loopback.cc | 135 DEFINE_int32(num_spatial_layers, 1, "Number of spatial layers to use."); 142 "Spatial layer to show or analyze. -1 to disable filtering.");
|
video_quality_test.h | 62 struct { // Spatial scalability.
|
/external/wpa_supplicant_8/src/drivers/ |
driver_common.c | 125 * An HT non-AP STA shall support all EQM rates for one spatial stream. 142 * A VHT non-AP STA shall support MCS 0-7 for one spatial stream.
|
/hardware/qcom/camera/msm8998/QCamera2/util/ |
QCameraFOVControl.cpp | 408 // WA for now until the QTI solution is in place writing the spatial alignment ready status [all...] |
/external/libvpx/libvpx/examples/ |
vp8_multi_resolution_encoder.c | 17 * This test also allows for settings temporal layers for each spatial layer. 18 * Different number of temporal layers per spatial stream may be used. 19 * Currently up to 3 temporal layers per spatial stream (encoder) are supported 51 /* Number of encoders (spatial resolutions) used in this test. */ 372 // Bitrates per spatial layer: overwrite default rates above. 377 // Temporal layers per spatial layers: overwrite default settings above. 462 // Set the number of threads per encode/spatial layer.
|
vp9_spatial_svc_encoder.c | 13 * VP9 encoding scheme based on spatial scalability for video applications 52 ARG_DEF("sl", "spatial-layers", 1, "number of spatial SVC layers"); 560 // Example pattern for spatial layers and 2 temporal layers used in the 683 // TODO(marpan): do spatial by re-writing superframe. 725 // and the buffer indices for each spatial layer of the current 739 // over all the spatial layers for the current superframe. [all...] |
/external/tensorflow/tensorflow/contrib/slim/python/slim/nets/ |
resnet_utils.py | 70 """Subsamples the input along the spatial dimensions. 160 output_stride, which is the ratio of the input to output spatial resolution. 178 ratio of input to output spatial resolution, which needs to be equal to
|
vgg.py | 91 spatial_squeeze: whether or not should squeeze the spatial dimensions of the 156 spatial_squeeze: whether or not should squeeze the spatial dimensions of the 221 spatial_squeeze: whether or not should squeeze the spatial dimensions of the
|
/external/tensorflow/tensorflow/core/kernels/ |
eigen_spatial_convolutions_test.cc | 209 // Apply a spatial convolution using a 4x4 kernel, valid padding, and a stride 264 // Apply a spatial convolution using a 4x4 kernel, valid padding, and a stride 323 // Apply a spatial convolution using a 4x4 kernel, valid padding, and a stride 375 // Apply a spatial convolution using a 3x3 kernel, valid padding, and a stride 423 // Apply a spatial convolution using a 1x1 kernel, valid padding, and a stride 474 // Apply a spatial convolution using a 3x3 kernel, valid padding, and a stride 522 // Apply a spatial convolution using a 3x3 kernel, valid padding 576 // Apply a spatial convolution using a 3x3 kernel, valid padding 630 // Apply a spatial convolution using a 3x3 kernel, valid padding [all...] |
/external/tensorflow/tensorflow/contrib/model_pruning/python/layers/ |
layers.py | 141 kernel_size: A sequence of N positive integers specifying the spatial 143 value for all spatial dimensions. 146 spatial dimensions. Specifying any `stride` value != 1 is incompatible 157 value for all spatial dimensions. Specifying any `rate` value != 1 is
|
/external/tensorflow/tensorflow/core/framework/ |
common_shape_fns.h | 40 // dimension. This is usually used to make sure that the spatial dimensions 70 // For 2D and 3D convolutions, the spatial dimensions are orthogonal, so the 71 // size and padding of each spatial dimension can be computed by calling 212 const std::vector<DimensionOrConstant>& spatial,
|
/frameworks/base/services/usb/java/com/android/server/usb/descriptors/ |
Usb20ASGeneral.java | 39 private int mChannelConfig; // 11:4 Describes the spatial location of the
|
/packages/inputmethods/LatinIME/native/jni/src/suggest/core/result/ |
suggestion_results.cpp | 93 AKLOGE("weight of language model vs spatial model: %f", mWeightOfLangModelVsSpatialModel);
|
/external/aac/libSACdec/src/ |
sac_dec_lib.cpp | 405 * \param pSsc spatial specific config handle. 668 /* signalize spatial decoder re-initalization */ 707 /* In config detection mode write spatial specific config parameters 722 /* In config detection mode write spatial specific config parameters 813 * \param pSsc spatial specific config handle. [all...] |
/external/libvpx/libvpx/vp8/common/ |
onyx.h | 103 * For spatial denoiser: noise_sensitivity controls the amount of 105 * Spatial denoiser invoked under !CONFIG_TEMPORAL_DENOISING.
|
/external/tensorflow/tensorflow/examples/android/jni/object_tracking/ |
image_data.h | 223 // Create the spatial derivatives for frame 1. 228 TimeLog("Created spatial derivatives");
|
/external/aac/libSACdec/include/ |
sac_dec_lib.h | 222 new header with a valid Spatial Specific Config and a 237 last spatial image before fading to the particular spatial 241 creates to fade from the last spatial scenario to the 246 creates to fade from the default spatial scenario (downmix) to 251 module starts fading from default to the current spatial
|
/external/libvpx/libvpx/test/ |
datarate_test.cc | [all...] |
/external/tensorflow/tensorflow/python/keras/_impl/keras/layers/ |
convolutional.py | 49 with the layer input over a single spatial (or temporal) dimension 160 """2D convolution layer (e.g. spatial convolution over images). 180 all spatial dimensions. 184 all spatial dimensions. 201 all spatial dimensions. 294 """3D convolution layer (e.g. spatial convolution over volumes). 315 all spatial dimensions. 317 specifying the strides of the convolution along each spatial 320 all spatial dimensions. 337 all spatial dimensions [all...] |
/external/tensorflow/tensorflow/contrib/bayesflow/python/ops/ |
layers_conv_variational.py | [all...] |
/external/webrtc/webrtc/modules/video_coding/codecs/vp9/ |
vp9_impl.cc | 78 // Use two spatial when screensharing with flexible mode. 114 // We check target_bitrate_bps of the 0th layer to see if the spatial layers 126 "spatial layers not implemented yet!"; 182 // For now, temporal layers only supported when having one spatial layer. 244 // libvpx currently supports only one or two spatial layers. 388 // Set QP-min/max per spatial and temporal layer. 432 // Register callback for getting each spatial layer. [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
vp9_denoiser.c | 258 // Shift for the second spatial layer. 455 // spatial layer was key frame. 584 // For SVC: only denoise at most 2 spatial (highest) layers. 586 // Denoise from one spatial layer below the top. 589 // Only denoise the top spatial layer.
|
/external/libvpx/libvpx/vpx/src/ |
svc_encodeframe.c | 223 if (strcmp("spatial-layers", option_name) == 0) { 413 svc_log(svc_ctx, SVC_LOG_ERROR, "spatial layers: invalid value: %d\n", 472 "spatial layers * temporal layers exceeds the maximum number of " 482 "1) spatial layer bitrates should sum up to target \n" 484 "a spatial layer \n");
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
conv_ops.cc | 218 "th spatial dimension had dilation ", 369 "th spatial dimension had dilation ", 433 // Mirror the filter in the spatial dimensions. 524 "th spatial dimension had dilation ", 559 // Each spatial entry has size in_depth * batch
|