/development/perftests/panorama/feature_mos/src/mosaic/ |
Mosaic.h | 34 The class Mosaic provides a simple interface to the panoramic mosaicing algorithm. The class allows passing in individual image frames to be stitched together, computes the alignment transformation between them, and then stitches and blends them together into a single panoramic output which can then be accessed as a single image. \ 47 while (<image frames are available>) 56 // Add to list of frames 93 * \param nframes Number of frames to pre-allocate; default value -1 will allocate each frame as it comes 95 * \param thresh_still Minimum number of pixels of translation detected between the new frame and the last frame before this frame is added to be mosaiced. For the low-res processing at 320x180 resolution input, we set this to 5 pixels. To reject no frames, set this to 0.0 (default value). 115 * After adding all frames, call this function to perform the final blending. 155 * Size of image frames making up mosaic 172 * Collection of frames that will make up mosaic. 174 MosaicFrame **frames; member in class:Mosaic 177 * Subset of frames that are considered as relevant [all...] |
/external/chromium_org/content/renderer/media/ |
webaudiosourceprovider_impl.cc | 117 const size_t frames = renderer_->Render(bus_wrapper_.get(), 0); local 118 if (frames < number_of_frames) 119 bus_wrapper_->ZeroFramesPartial(frames, number_of_frames - frames);
|
renderer_webaudiodevice_impl.cc | 92 dest->frames()); 95 return dest->frames();
|
/external/chromium_org/media/base/ |
fake_audio_render_callback.cc | 27 int number_of_frames = audio_bus->frames();
|
sinc_resampler_perftest.cc | 21 static void DoNothing(int frames, float* destination) {}
|
audio_renderer_mixer_input.cc | 102 // AudioConverter expects unfilled frames to be zeroed. 103 if (frames_filled < audio_bus->frames()) { 105 frames_filled, audio_bus->frames() - frames_filled);
|
audio_timestamp_helper_unittest.cc | 19 // Adds frames to the helper and returns the current timestamp in 21 int64 AddFrames(int frames) { 22 helper_.AddFrames(frames); 58 // Verify that adding frames one frame at a time matches the timestamp 59 // returned if the same number of frames are added all at once. 99 // Verify that the same number of frames is returned up 109 // Add frames to the helper so negative frame counts can be tested.
|
/external/chromium_org/media/filters/ |
audio_file_reader.cc | 123 // Read until we hit EOF or we've read the requested number of frames. 128 while (current_frame < audio_bus->frames() && continue_decoding && 130 // Make a shallow copy of packet so we can slide packet.data as frames are 156 // Determine the number of sample-frames we just decoded. Check overflow. 186 if (current_frame + frames_read > audio_bus->frames()) { 188 frames_read = audio_bus->frames() - current_frame; 219 // Zero any remaining frames. 221 current_frame, audio_bus->frames() - current_frame); 223 // Returns the actual number of sample-frames decoded. 232 // |duration| has been calculated from an exact number of sample-frames [all...] |
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/network/ |
ResourceWebSocketFrameView.js | 104 var frames = this.resource.frames(); 105 for (var i = frames.length - 1; i >= 0; i--) { 106 this.appendFrame(frames[i]);
|
/external/chromium_org/third_party/WebKit/public/web/ |
WebPageSerializer.h | 45 // Get html data by serializing all frames of current page with lists 55 // Serializes all the frames from the WebView, retrieves the page's 73 // This function will find out all frames and serialize them to HTML data. 82 // serialize all sub frames of the specified frame or not. 90 // contain all saved auxiliary files included all sub frames and resources. 100 // and sub-frames. Returns true if all resources were retrieved 105 WebVector<WebURL>* frames);
|
/external/chromium_org/third_party/webrtc/modules/video_processing/main/test/unit_test/ |
readYUV420file.m | 15 % calculate number of frames 20 % calculate number of frames
|
/external/kernel-headers/original/uapi/linux/can/ |
bcm.h | 54 * @count: number of frames to send before changing interval. 55 * @ival1: interval for the first @count frames. 56 * @ival2: interval for the following frames. 57 * @can_id: CAN ID of frames to be sent or received. 58 * @nframes: number of frames appended to the message head. 59 * @frames: array of CAN frames. 68 struct can_frame frames[0]; member in struct:bcm_msg_head
|
/packages/apps/Camera/jni/feature_mos/src/mosaic/ |
Mosaic.h | 34 The class Mosaic provides a simple interface to the panoramic mosaicing algorithm. The class allows passing in individual image frames to be stitched together, computes the alignment transformation between them, and then stitches and blends them together into a single panoramic output which can then be accessed as a single image. \ 47 while (<image frames are available>) 56 // Add to list of frames 93 * \param nframes Number of frames to pre-allocate; default value -1 will allocate each frame as it comes 95 * \param thresh_still Minimum number of pixels of translation detected between the new frame and the last frame before this frame is added to be mosaiced. For the low-res processing at 320x180 resolution input, we set this to 5 pixels. To reject no frames, set this to 0.0 (default value). 115 * After adding all frames, call this function to perform the final blending. 155 * Size of image frames making up mosaic 172 * Collection of frames that will make up mosaic. 174 MosaicFrame **frames; member in class:Mosaic 177 * Subset of frames that are considered as relevant [all...] |
/packages/apps/LegacyCamera/jni/feature_mos/src/mosaic/ |
Mosaic.h | 34 The class Mosaic provides a simple interface to the panoramic mosaicing algorithm. The class allows passing in individual image frames to be stitched together, computes the alignment transformation between them, and then stitches and blends them together into a single panoramic output which can then be accessed as a single image. \ 47 while (<image frames are available>) 56 // Add to list of frames 93 * \param nframes Number of frames to pre-allocate; default value -1 will allocate each frame as it comes 95 * \param thresh_still Minimum number of pixels of translation detected between the new frame and the last frame before this frame is added to be mosaiced. For the low-res processing at 320x180 resolution input, we set this to 5 pixels. To reject no frames, set this to 0.0 (default value). 115 * After adding all frames, call this function to perform the final blending. 155 * Size of image frames making up mosaic 172 * Collection of frames that will make up mosaic. 174 MosaicFrame **frames; member in class:Mosaic 177 * Subset of frames that are considered as relevant [all...] |
/prebuilts/ndk/9/platforms/android-19/arch-arm64/usr/include/linux/can/ |
bcm.h | 32 struct can_frame frames[0]; member in struct:bcm_msg_head
|
/prebuilts/ndk/9/platforms/android-19/arch-mips64/usr/include/linux/can/ |
bcm.h | 32 struct can_frame frames[0]; member in struct:bcm_msg_head
|
/prebuilts/ndk/9/platforms/android-19/arch-x86_64/usr/include/linux/can/ |
bcm.h | 32 struct can_frame frames[0]; member in struct:bcm_msg_head
|
/system/media/audio_utils/include/audio_utils/ |
sndfile.h | 35 sf_count_t frames; member in struct:__anon81379 63 // Read interleaved frames and return actual number of frames read 68 // Write interleaved frames and return actual number of frames written
|
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/timeline/ |
TimelineFrameModel.js | 47 frames: function() 77 var frames = this._frames; 78 var firstFrame = insertionIndexForObjectInListSortedByFunction(startTime, frames, compareEndTime); 79 var lastFrame = insertionIndexForObjectInListSortedByFunction(endTime, frames, compareStartTime); 80 return frames.slice(firstFrame, lastFrame); 115 // - only show frames that either did not wait for the main thread frame or had one committed. 477 * @param {!Array.<!WebInspector.TimelineFrame>} frames 479 WebInspector.FrameStatistics = function(frames) 481 this.frameCount = frames.length; 485 this.startOffset = frames[0].startTimeOffset [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/ |
vp9_temporal_filter.c | 203 YV12_BUFFER_CONFIG *f = cpi->frames[alt_ref_index]; 219 // Source frames are extended to 16 pixels. This is different than 220 // L/A/G reference frames that have a border of 32 (VP9ENCBORDERINPIXELS) 249 if (cpi->frames[frame] == NULL) 260 cpi->frames[alt_ref_index]->y_buffer + mb_y_offset, 261 cpi->frames[frame]->y_buffer + mb_y_offset, 262 cpi->frames[frame]->y_stride); 274 cpi->frames[frame]->y_buffer + mb_y_offset, 275 cpi->frames[frame]->u_buffer + mb_uv_offset, 276 cpi->frames[frame]->v_buffer + mb_uv_offset [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
vp9_temporal_filter.c | 206 YV12_BUFFER_CONFIG *f = cpi->frames[alt_ref_index]; 223 // Source frames are extended to 16 pixels. This is different than 224 // L/A/G reference frames that have a border of 32 (VP9ENCBORDERINPIXELS) 253 if (cpi->frames[frame] == NULL) 270 cpi->frames[alt_ref_index]->y_buffer + mb_y_offset, 271 cpi->frames[frame]->y_buffer + mb_y_offset, 272 cpi->frames[frame]->y_stride); 285 cpi->frames[frame]->y_buffer + mb_y_offset, 286 cpi->frames[frame]->u_buffer + mb_uv_offset, 287 cpi->frames[frame]->v_buffer + mb_uv_offset [all...] |
/frameworks/native/opengl/tests/hwc/ |
hwcStress.cpp | 198 static vector <vector <sp<GraphicBuffer> > > frames; variable 395 // Regenerate a new set of test frames when this pass is 412 list = hwcTestCreateLayerList(testRandMod(frames.size()) + 1); 418 // Prandomly select a subset of frames to be used by this pass. 420 selectedFrames = vectorRandSelect(frames, list->numHwLayers); 541 * Initialize Frames 544 * named frames. The graphic buffers are contained within a vector of 561 frames.clear(); 562 frames.resize(rows); 565 // All frames within a row have to have the same format an [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/ |
vp9_temporal_filter.c | 206 YV12_BUFFER_CONFIG *f = cpi->frames[alt_ref_index]; 223 // Source frames are extended to 16 pixels. This is different than 224 // L/A/G reference frames that have a border of 32 (VP9ENCBORDERINPIXELS) 253 if (cpi->frames[frame] == NULL) 270 cpi->frames[alt_ref_index]->y_buffer + mb_y_offset, 271 cpi->frames[frame]->y_buffer + mb_y_offset, 272 cpi->frames[frame]->y_stride); 285 cpi->frames[frame]->y_buffer + mb_y_offset, 286 cpi->frames[frame]->u_buffer + mb_uv_offset, 287 cpi->frames[frame]->v_buffer + mb_uv_offset [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/test/ |
superframe_test.cc | 51 const int frames = (marker & 0x7) + 1; local 53 const unsigned int index_sz = 2 + mag * frames; 72 // Make sure we do a few frames after the last SF
|
/external/compiler-rt/lib/sanitizer_common/ |
sanitizer_symbolizer_win.cc | 29 uptr SymbolizePC(uptr addr, AddressInfo *frames, uptr max_frames) { 59 AddressInfo *info = &frames[0];
|