/frameworks/av/media/libaaudio/src/flowgraph/ |
MonoToMultiConverter.h | 34 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
|
SourceFloat.h | 31 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
|
SourceI16.h | 31 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
|
SourceI24.h | 31 int32_t onProcess(int64_t framePosition, int32_t numFrames) override;
|
SourceI16.cpp | 33 int32_t SourceI16::onProcess(int64_t framePosition, int32_t numFrames) { 38 int32_t framesToProcess = std::min(numFrames, framesLeft);
|
SinkI24.cpp | 33 int32_t SinkI24::read(void *data, int32_t numFrames) { 37 int32_t framesLeft = numFrames; 65 return numFrames - framesLeft;
|
/external/webrtc/webrtc/modules/video_processing/test/ |
readYUV420file.m | 21 numFrames = floor(fileLen/frameSizeBytes); 23 Y=uint8(zeros(height,width,numFrames)); 24 U=uint8(zeros(height/2,width/2,numFrames)); 25 V=uint8(zeros(height/2,width/2,numFrames)); 29 for k=1:numFrames
|
/frameworks/av/media/libeffects/downmix/ |
EffectDownmix.c | 338 size_t numFrames = outBuffer->frameCount; 348 while (numFrames) { 353 numFrames--; 356 while (numFrames) { 361 numFrames--; 370 downmixInputChannelMask, pSrc, pDst, numFrames, accumulate)) { 380 Downmix_foldFromQuad(pSrc, pDst, numFrames, accumulate); 384 Downmix_foldFrom5Point1(pSrc, pDst, numFrames, accumulate); 387 Downmix_foldFrom7Point1(pSrc, pDst, numFrames, accumulate); 391 downmixInputChannelMask, pSrc, pDst, numFrames, accumulate)) [all...] |
/frameworks/av/media/libaaudio/tests/ |
test_atomic_fifo.cpp | 164 fifo_frames_t bothAvailable = wrappingBuffer.numFrames[0] + wrappingBuffer.numFrames[1]; 171 bothAvailable = wrappingBuffer.numFrames[0] + wrappingBuffer.numFrames[1]; 176 EXPECT_GE(wrappingBuffer.numFrames[0], 0); 177 EXPECT_LE(wrappingBuffer.numFrames[0], capacity); 178 EXPECT_GE(wrappingBuffer.numFrames[1], 0); 179 EXPECT_LE(wrappingBuffer.numFrames[1], capacity); 183 if (wrappingBuffer.numFrames[0]) { 186 + mFifoBuffer.convertFramesToBytes(wrappingBuffer.numFrames[0]) - 1 [all...] |
/cts/apps/CtsVerifier/src/com/android/cts/verifier/audio/audiolib/ |
WaveScopeView.java | 55 public void setPCM16Buff(short[] smpl16Buff, int numChans, int numFrames) { 60 mNumFrames = numFrames; 67 public void setPCMFloatBuff(float[] smplFloatBuff, int numChans, int numFrames) { 72 mNumFrames = numFrames; 111 * @param numFrames The number of FRAMES in the specified sample block. 116 private void drawChannel16(Canvas cvs, short[] samples, int numFrames, int numChans, 122 for (int frame = 0; frame < numFrames; frame++) { 140 * @param numFrames The number of FRAMES in the specified sample block. 145 private void drawChannelFloat(Canvas cvs, float[] samples, int numFrames, int numChans, 151 for (int frame = 0; frame < numFrames; frame++) [all...] |
WaveTableFloatFiller.java | 65 public int fill(float[] buffer, int numFrames, int numChans) { 68 for (int frameIndex = 0; frameIndex < numFrames; frameIndex++) { 87 return numFrames;
|
/developers/build/prebuilts/gradle/MidiScope/Application/src/main/java/com/example/android/common/midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/developers/build/prebuilts/gradle/MidiSynth/Application/src/main/java/com/example/android/common/midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/developers/samples/android/common/src/java/com/example/android/common/midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/development/samples/browseable/MidiScope/src/com.example.android.common.midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/development/samples/browseable/MidiSynth/src/com.example.android.common.midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/frameworks/av/media/libaaudio/src/client/ |
AAudioFlowGraph.h | 46 void process(const void *source, void *destination, int32_t numFrames); 53 void setRampLengthInFrames(int32_t numFrames);
|
AudioStreamInternalCapture.cpp | 60 aaudio_result_t AudioStreamInternalCapture::read(void *buffer, int32_t numFrames, 63 return processData(buffer, numFrames, timeoutNanoseconds); 67 aaudio_result_t AudioStreamInternalCapture::processDataNow(void *buffer, int32_t numFrames, 115 //ALOGD("AudioStreamInternalCapture::processDataNow() - readNowWithConversion(%d)", numFrames); 116 int32_t framesProcessed = readNowWithConversion(buffer, numFrames); 118 // numFrames, framesProcessed); 157 int32_t numFrames) { 159 // buffer, numFrames); 162 int32_t framesLeft = numFrames; 169 const int32_t framesAvailable = wrappingBuffer.numFrames[partIndex] [all...] |
AAudioFlowGraph.cpp | 102 void AAudioFlowGraph::process(const void *source, void *destination, int32_t numFrames) { 103 mSource->setData(source, numFrames); 104 mSink->read(destination, numFrames); 114 void AAudioFlowGraph::setRampLengthInFrames(int32_t numFrames) { 115 mVolumeRamp->setLengthInFrames(numFrames);
|
AudioStreamInternalPlay.cpp | 55 int32_t numFrames = kRampMSec * getSampleRate() / AAUDIO_MILLIS_PER_SECOND; 56 mFlowGraph.setRampLengthInFrames(numFrames); 109 aaudio_result_t AudioStreamInternalPlay::write(const void *buffer, int32_t numFrames, 111 return processData((void *)buffer, numFrames, timeoutNanoseconds); 115 aaudio_result_t AudioStreamInternalPlay::processDataNow(void *buffer, int32_t numFrames, 162 //ALOGD("AudioStreamInternal::processDataNow() - writeNowWithConversion(%d)", numFrames); 163 int32_t framesWritten = writeNowWithConversion(buffer, numFrames); 165 // numFrames, framesWritten); 210 int32_t numFrames) { 213 int32_t framesLeft = numFrames; [all...] |
/external/mp4parser/isoparser/src/main/java/com/googlecode/mp4parser/authoring/tracks/ |
SilenceTrackImpl.java | 25 long numFrames = getTrackMetaData().getTimescale() * ms / 1000 / 1024; 26 long standZeit = getTrackMetaData().getTimescale() * ms / numFrames / 1000; 27 entry = new TimeToSampleBox.Entry(numFrames, standZeit); 30 while (numFrames-- > 0) {
|
/frameworks/av/media/libaaudio/src/legacy/ |
AudioStreamLegacy.cpp | 54 int32_t numFrames) { 58 incrementFramesRead(numFrames); 59 finalAudioData = (void *) maybeConvertDeviceData(buffer, numFrames); 63 aaudio_data_callback_result_t callbackResult = maybeCallDataCallback(finalAudioData, numFrames); 68 incrementFramesWritten(numFrames); 75 int32_t numFrames = numBytes / getBytesPerDeviceFrame(); 76 return (int32_t) callDataCallbackFrames(buffer, numFrames);
|
/frameworks/av/media/libaaudio/examples/utils/ |
AAudioSimplePlayer.h | 179 const int numFrames = 32; 180 float zeros[numFrames * samplesPerFrame]; 182 aaudio_result_t result = numFrames; 183 while (result == numFrames) { 184 result = AAudioStream_write(mStream, zeros, numFrames, 0); 312 int32_t numFrames 353 if (numFrames > sineData->maxNumFrames) { 354 sineData->maxNumFrames = numFrames; 356 if (numFrames < sineData->minNumFrames) { 357 sineData->minNumFrames = numFrames; [all...] |
/cts/tests/openglperf2/jni/primitive/ |
GLPrimitive.cpp | 43 JNIEnv* env, jclass /*clazz*/, jint workload, jint numFrames, jdoubleArray frameTimes) { 56 numFrames /= Renderer::OFFSCREEN_INNER_FRAMES; 60 for (int i = 0; i < numFrames && success; i++) {
|
/frameworks/av/services/oboeservice/ |
AAudioMixer.cpp | 89 fifo_frames_t framesAvailableFromPart = wrappingBuffer.numFrames[partIndex]; 111 void AAudioMixer::mixPart(float *destination, float *source, int32_t numFrames) { 112 int32_t numSamples = numFrames * mSamplesPerFrame;
|