/external/chromium_org/media/audio/cras/ |
cras_unified.cc | 258 unsigned int frames, 263 return me->DispatchCallback(frames, 281 uint32 CrasUnifiedStream::DispatchCallback(size_t frames, 288 return WriteAudio(frames, output_samples, output_ts); 299 uint32 CrasUnifiedStream::WriteAudio(size_t frames, 302 DCHECK_EQ(frames, static_cast<size_t>(output_bus_->frames()));
|
/external/compiler-rt/lib/sanitizer_common/ |
sanitizer_symbolizer_libbacktrace.cc | 87 AddressInfo *frames; member in struct:__sanitizer::__anon24250::SymbolizeCodeData 100 AddressInfo *info = &cdata->frames[cdata->n_frames++]; 118 AddressInfo *info = &cdata->frames[0]; 151 uptr LibbacktraceSymbolizer::SymbolizeCode(uptr addr, AddressInfo *frames, 156 data.frames = frames; 182 uptr LibbacktraceSymbolizer::SymbolizeCode(uptr addr, AddressInfo *frames,
|
/libcore/dalvik/src/main/java/dalvik/system/profiler/ |
HprofData.java | 181 StringBuilder frames = new StringBuilder(); local 183 frames.append('\n'); 185 frames.append("\t at "); 186 frames.append(stackFrame); 187 frames.append('\n'); 190 frames.append("<empty>"); 194 + ", frames=" + frames + "]";
|
/external/chromium_org/third_party/webrtc/modules/audio_coding/codecs/opus/ |
opus_interface.c | 21 /* The format allows up to 120 ms frames. Since we don't control the other 464 int frames, samples; local 465 frames = opus_packet_get_nb_frames(payload, payload_length_bytes); 466 if (frames < 0) { 470 samples = frames * opus_packet_get_samples_per_frame(payload, 48000); 495 int frames, channels, payload_length_ms; local 516 frames = 1; 520 frames = 2; 524 frames = 3; 543 if (frame_data[0][0] & (0x80 >> ((n + 1) * (frames + 1) - 1)) [all...] |
/frameworks/wilhelm/tools/permute/ |
permute.c | 75 /** Split the specified range of frames, using the allowed budget of segments. 202 switch (sfinfo_in.frames) { 205 fprintf(stderr, "%s: unsupported frames %d\n", path_in, (int) sfinfo_in.frames); 212 double durationSeconds = (double) sfinfo_in.frames / (double) sfinfo_in.samplerate; 224 used = split(&s, 0, sfinfo_in.frames, s.mSegmentMax); 241 void *ptr = malloc(sfinfo_in.frames * frameSizeRead); 244 count = sf_readf_short(sf_in, ptr, sfinfo_in.frames); 245 if (count != sfinfo_in.frames) { 246 fprintf(stderr, "%s: expected to read %d frames but actually read %d frames\n", path_in [all...] |
/cts/apps/CameraITS/tests/sensor_fusion/ |
test_sensor_fusion.py | 34 # Capture 210 QVGA frames (which is 7s at 30fps) 77 # timestamps are in the "events" dictionary, and "frames" is a list of 80 events, frames = collect_data() 82 events, frames = load_data() 100 # adjacent frames. 101 cam_rots = get_cam_rotations(frames) 191 pylab.ylabel("Angular displacement between adjacent camera frames (deg)") 214 # Get the window of gyro samples within the current pair of frames. 247 def get_cam_rotations(frames): 248 """Get the rotations of the camera between each pair of frames [all...] |
/external/chromium_org/content/browser/frame_host/ |
render_frame_proxy_host.cc | 37 RoutingIDFrameProxyMap* frames = g_routing_id_frame_proxy_map.Pointer(); local 38 RoutingIDFrameProxyMap::iterator it = frames->find( 40 return it == frames->end() ? NULL : it->second;
|
/external/chromium_org/content/renderer/media/ |
webaudio_capturer_source.cc | 103 int available = fifo_->max_frames() - fifo_->frames(); 123 while (fifo_->frames() >= capture_frames) { 126 capture_bus_->ToInterleaved(capture_bus_->frames(),
|
/external/chromium_org/third_party/WebKit/Source/platform/exported/ |
WebImageSkia.cpp | 60 // Frames are arranged by decreasing size, then decreasing bit depth. 75 break; // No more frames that are large enough. 108 // Frames are arranged by decreasing size, then decreasing bit depth. 113 Vector<WebImage> frames; local 126 frames.append(WebImage(image->bitmap())); 129 return frames;
|
/external/chromium_org/tools/tcmalloc/ |
print-live-objects.py | 36 "frames": matches.group(3).strip().split(" ")} 44 for frame in trace["frames"]: 62 frames = [] 63 for frame in trace["frames"]: 64 frames.append(table[frame]) 65 trace["frames"] = frames 85 for frame in trace["frames"]:
|
/external/javassist/src/main/javassist/bytecode/analysis/ |
Analyzer.java | 74 * Frame[] frames = analyzer.analyze(clazz, method); 75 * frames[0].getLocal(0).getCtClass(); // returns clazz; 76 * frames[0].getLocal(1).getCtClass(); // returns java.lang.String 77 * frames[1].peek(); // returns Type.INTEGER 78 * frames[27].peek().getCtClass(); // returns java.lang.Number 89 private Frame[] frames; field in class:Analyzer 139 frames = new Frame[codeLength]; 140 frames[iter.lookAhead()] = firstFrame(method, maxLocals, maxStack); 146 return frames; 173 Frame frame = frames[pos].copy() [all...] |
FramePrinter.java | 85 Frame[] frames; 87 frames = (new Analyzer()).analyze(method.getDeclaringClass(), info); 106 Frame frame = frames[pos];
|
/bionic/libc/kernel/uapi/linux/can/ |
bcm.h | 32 struct can_frame frames[0]; member in struct:bcm_msg_head
|
/cts/tests/tests/textureview/src/android/textureview/cts/ |
GLProducerThread.java | 59 int frames, int delayMs, Semaphore semaphore) { 61 mFrames = frames; 68 GLProducerThread(SurfaceTexture surfaceTexture, GLRenderer renderer, int frames, int delayMs, 70 this(surfaceTexture, renderer, null, frames, delayMs, semaphore);
|
/development/ndk/platforms/android-L/include/linux/can/ |
bcm.h | 32 struct can_frame frames[0]; member in struct:bcm_msg_head
|
/development/perftests/panorama/feature_mos/src/mosaic/ |
Mosaic.h | 34 The class Mosaic provides a simple interface to the panoramic mosaicing algorithm. The class allows passing in individual image frames to be stitched together, computes the alignment transformation between them, and then stitches and blends them together into a single panoramic output which can then be accessed as a single image. \ 47 while (<image frames are available>) 56 // Add to list of frames 93 * \param nframes Number of frames to pre-allocate; default value -1 will allocate each frame as it comes 95 * \param thresh_still Minimum number of pixels of translation detected between the new frame and the last frame before this frame is added to be mosaiced. For the low-res processing at 320x180 resolution input, we set this to 5 pixels. To reject no frames, set this to 0.0 (default value). 115 * After adding all frames, call this function to perform the final blending. 155 * Size of image frames making up mosaic 172 * Collection of frames that will make up mosaic. 174 MosaicFrame **frames; member in class:Mosaic 177 * Subset of frames that are considered as relevant [all...] |
/external/chromium_org/media/base/ |
fake_audio_render_callback.cc | 27 int number_of_frames = audio_bus->frames();
|
sinc_resampler_perftest.cc | 21 static void DoNothing(int frames, float* destination) {}
|
/external/chromium_org/media/blink/ |
webaudiosourceprovider_impl.cc | 117 const int frames = renderer_->Render(bus_wrapper_.get(), 0); local 118 if (frames < static_cast<int>(number_of_frames)) { 120 frames, 121 static_cast<int>(number_of_frames - frames));
|
/external/chromium_org/media/filters/ |
audio_renderer_algorithm.h | 42 // Tries to fill |requested_frames| frames into |dest| with possibly scaled 48 // Returns the number of frames copied into |dest|. 61 // Returns the capacity of |audio_buffer_| in frames. 67 // Returns the number of frames left in |audio_buffer_|, which may be larger 70 int frames_buffered() { return audio_buffer_.frames(); } 79 // |target_block_| can be extracted from the available frames. 82 // Read a maximum of |requested_frames| frames from |wsola_output_|. Returns 83 // number of frames actually read. 87 // Fill |dest| with frames from |audio_buffer_| starting from frame 91 // for negative indices. This might happen for few first frames. This metho [all...] |
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/network/ |
ResourceWebSocketFrameView.js | 88 var frames = this._request.frames(); 89 for (var i = frames.length - 1; i >= 0; --i) 90 this._dataGrid.insertChild(new WebInspector.ResourceWebSocketFrameNode(frames[i]));
|
/external/chromium_org/third_party/WebKit/public/web/ |
WebPageSerializer.h | 46 // Get html data by serializing all frames of current page with lists 56 // Serializes all the frames from the WebView, retrieves the page's 74 // This function will find out all frames and serialize them to HTML data. 83 // serialize all sub frames of the specified frame or not. 91 // contain all saved auxiliary files included all sub frames and resources. 101 // and sub-frames. Returns true if all resources were retrieved 106 WebVector<WebURL>* frames);
|
/external/chromium_org/third_party/decklink/mac/include/ |
DeckLinkAPITypes.h | 81 virtual HRESULT GetComponents (/* out */ uint8_t *hours, /* out */ uint8_t *minutes, /* out */ uint8_t *seconds, /* out */ uint8_t *frames) = 0;
|
/external/chromium_org/third_party/webrtc/modules/video_processing/main/test/unit_test/ |
readYUV420file.m | 15 % calculate number of frames 20 % calculate number of frames
|
/external/kernel-headers/original/uapi/linux/can/ |
bcm.h | 54 * @count: number of frames to send before changing interval. 55 * @ival1: interval for the first @count frames. 56 * @ival2: interval for the following frames. 57 * @can_id: CAN ID of frames to be sent or received. 58 * @nframes: number of frames appended to the message head. 59 * @frames: array of CAN frames. 68 struct can_frame frames[0]; member in struct:bcm_msg_head
|