HomeSort by relevance Sort by last modified time
    Searched refs:numFrames (Results 26 - 50 of 137) sorted by null

12 3 4 5 6

  /frameworks/av/media/libaaudio/src/flowgraph/
MonoToMultiConverter.h 34 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
SourceFloat.h 31 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
SourceI16.h 31 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
SourceI24.h 31 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
SourceI16.cpp 33 int32_t SourceI16::onProcess(int64_t framePosition, int32_t numFrames) {
38 int32_t framesToProcess = std::min(numFrames, framesLeft);
SinkI24.cpp 33 int32_t SinkI24::read(void *data, int32_t numFrames) {
37 int32_t framesLeft = numFrames;
65 return numFrames - framesLeft;
  /external/webrtc/webrtc/modules/video_processing/test/
readYUV420file.m 21 numFrames = floor(fileLen/frameSizeBytes);
23 Y=uint8(zeros(height,width,numFrames));
24 U=uint8(zeros(height/2,width/2,numFrames));
25 V=uint8(zeros(height/2,width/2,numFrames));
29 for k=1:numFrames
  /frameworks/av/media/libeffects/downmix/
EffectDownmix.c 338 size_t numFrames = outBuffer->frameCount;
348 while (numFrames) {
353 numFrames--;
356 while (numFrames) {
361 numFrames--;
370 downmixInputChannelMask, pSrc, pDst, numFrames, accumulate)) {
380 Downmix_foldFromQuad(pSrc, pDst, numFrames, accumulate);
384 Downmix_foldFrom5Point1(pSrc, pDst, numFrames, accumulate);
387 Downmix_foldFrom7Point1(pSrc, pDst, numFrames, accumulate);
391 downmixInputChannelMask, pSrc, pDst, numFrames, accumulate))
    [all...]
  /frameworks/av/media/libaaudio/tests/
test_atomic_fifo.cpp 164 fifo_frames_t bothAvailable = wrappingBuffer.numFrames[0] + wrappingBuffer.numFrames[1];
171 bothAvailable = wrappingBuffer.numFrames[0] + wrappingBuffer.numFrames[1];
176 EXPECT_GE(wrappingBuffer.numFrames[0], 0);
177 EXPECT_LE(wrappingBuffer.numFrames[0], capacity);
178 EXPECT_GE(wrappingBuffer.numFrames[1], 0);
179 EXPECT_LE(wrappingBuffer.numFrames[1], capacity);
183 if (wrappingBuffer.numFrames[0]) {
186 + mFifoBuffer.convertFramesToBytes(wrappingBuffer.numFrames[0]) - 1
    [all...]
  /cts/apps/CtsVerifier/src/com/android/cts/verifier/audio/audiolib/
WaveScopeView.java 55 public void setPCM16Buff(short[] smpl16Buff, int numChans, int numFrames) {
60 mNumFrames = numFrames;
67 public void setPCMFloatBuff(float[] smplFloatBuff, int numChans, int numFrames) {
72 mNumFrames = numFrames;
111 * @param numFrames The number of FRAMES in the specified sample block.
116 private void drawChannel16(Canvas cvs, short[] samples, int numFrames, int numChans,
122 for (int frame = 0; frame < numFrames; frame++) {
140 * @param numFrames The number of FRAMES in the specified sample block.
145 private void drawChannelFloat(Canvas cvs, float[] samples, int numFrames, int numChans,
151 for (int frame = 0; frame < numFrames; frame++)
    [all...]
WaveTableFloatFiller.java 65 public int fill(float[] buffer, int numFrames, int numChans) {
68 for (int frameIndex = 0; frameIndex < numFrames; frameIndex++) {
87 return numFrames;
  /developers/build/prebuilts/gradle/MidiScope/Application/src/main/java/com/example/android/common/midi/synth/
SynthVoice.java 51 int numFrames = outputBuffer.length / samplesPerFrame;
52 for (int i = 0; i < numFrames; i++) {
  /developers/build/prebuilts/gradle/MidiSynth/Application/src/main/java/com/example/android/common/midi/synth/
SynthVoice.java 51 int numFrames = outputBuffer.length / samplesPerFrame;
52 for (int i = 0; i < numFrames; i++) {
  /developers/samples/android/common/src/java/com/example/android/common/midi/synth/
SynthVoice.java 51 int numFrames = outputBuffer.length / samplesPerFrame;
52 for (int i = 0; i < numFrames; i++) {
  /development/samples/browseable/MidiScope/src/com.example.android.common.midi/synth/
SynthVoice.java 51 int numFrames = outputBuffer.length / samplesPerFrame;
52 for (int i = 0; i < numFrames; i++) {
  /development/samples/browseable/MidiSynth/src/com.example.android.common.midi/synth/
SynthVoice.java 51 int numFrames = outputBuffer.length / samplesPerFrame;
52 for (int i = 0; i < numFrames; i++) {
  /frameworks/av/media/libaaudio/src/client/
AAudioFlowGraph.h 46 void process(const void *source, void *destination, int32_t numFrames);
53 void setRampLengthInFrames(int32_t numFrames);
AudioStreamInternalCapture.cpp 60 aaudio_result_t AudioStreamInternalCapture::read(void *buffer, int32_t numFrames,
63 return processData(buffer, numFrames, timeoutNanoseconds);
67 aaudio_result_t AudioStreamInternalCapture::processDataNow(void *buffer, int32_t numFrames,
115 //ALOGD("AudioStreamInternalCapture::processDataNow() - readNowWithConversion(%d)", numFrames);
116 int32_t framesProcessed = readNowWithConversion(buffer, numFrames);
118 // numFrames, framesProcessed);
157 int32_t numFrames) {
159 // buffer, numFrames);
162 int32_t framesLeft = numFrames;
169 const int32_t framesAvailable = wrappingBuffer.numFrames[partIndex]
    [all...]
AAudioFlowGraph.cpp 102 void AAudioFlowGraph::process(const void *source, void *destination, int32_t numFrames) {
103 mSource->setData(source, numFrames);
104 mSink->read(destination, numFrames);
114 void AAudioFlowGraph::setRampLengthInFrames(int32_t numFrames) {
115 mVolumeRamp->setLengthInFrames(numFrames);
AudioStreamInternalPlay.cpp 55 int32_t numFrames = kRampMSec * getSampleRate() / AAUDIO_MILLIS_PER_SECOND;
56 mFlowGraph.setRampLengthInFrames(numFrames);
109 aaudio_result_t AudioStreamInternalPlay::write(const void *buffer, int32_t numFrames,
111 return processData((void *)buffer, numFrames, timeoutNanoseconds);
115 aaudio_result_t AudioStreamInternalPlay::processDataNow(void *buffer, int32_t numFrames,
162 //ALOGD("AudioStreamInternal::processDataNow() - writeNowWithConversion(%d)", numFrames);
163 int32_t framesWritten = writeNowWithConversion(buffer, numFrames);
165 // numFrames, framesWritten);
210 int32_t numFrames) {
213 int32_t framesLeft = numFrames;
    [all...]
  /external/mp4parser/isoparser/src/main/java/com/googlecode/mp4parser/authoring/tracks/
SilenceTrackImpl.java 25 long numFrames = getTrackMetaData().getTimescale() * ms / 1000 / 1024;
26 long standZeit = getTrackMetaData().getTimescale() * ms / numFrames / 1000;
27 entry = new TimeToSampleBox.Entry(numFrames, standZeit);
30 while (numFrames-- > 0) {
  /frameworks/av/media/libaaudio/src/legacy/
AudioStreamLegacy.cpp 54 int32_t numFrames) {
58 incrementFramesRead(numFrames);
59 finalAudioData = (void *) maybeConvertDeviceData(buffer, numFrames);
63 aaudio_data_callback_result_t callbackResult = maybeCallDataCallback(finalAudioData, numFrames);
68 incrementFramesWritten(numFrames);
75 int32_t numFrames = numBytes / getBytesPerDeviceFrame();
76 return (int32_t) callDataCallbackFrames(buffer, numFrames);
  /frameworks/av/media/libaaudio/examples/utils/
AAudioSimplePlayer.h 179 const int numFrames = 32;
180 float zeros[numFrames * samplesPerFrame];
182 aaudio_result_t result = numFrames;
183 while (result == numFrames) {
184 result = AAudioStream_write(mStream, zeros, numFrames, 0);
312 int32_t numFrames
353 if (numFrames > sineData->maxNumFrames) {
354 sineData->maxNumFrames = numFrames;
356 if (numFrames < sineData->minNumFrames) {
357 sineData->minNumFrames = numFrames;
    [all...]
  /cts/tests/openglperf2/jni/primitive/
GLPrimitive.cpp 43 JNIEnv* env, jclass /*clazz*/, jint workload, jint numFrames, jdoubleArray frameTimes) {
56 numFrames /= Renderer::OFFSCREEN_INNER_FRAMES;
60 for (int i = 0; i < numFrames && success; i++) {
  /frameworks/av/services/oboeservice/
AAudioMixer.cpp 89 fifo_frames_t framesAvailableFromPart = wrappingBuffer.numFrames[partIndex];
111 void AAudioMixer::mixPart(float *destination, float *source, int32_t numFrames) {
112 int32_t numSamples = numFrames * mSamplesPerFrame;

Completed in 576 milliseconds

12 3 4 5 6