/cts/tests/tests/media/libndkaudio/ |
AudioSink.h | 27 virtual bool put(float * buff, int numFrames, int numChannels) =0;
|
AudioSource.h | 37 virtual int getData(long time, float * buff, int numFrames,
|
WaveTableOscillator.cpp | 47 int WaveTableOscillator::getData(long time, float* outBuff, int numFrames, 60 for (int dstIndex = 0; dstIndex < numFrames; ++dstIndex) { 78 for (int frameIndex = 0; frameIndex < numFrames; frameIndex++) { 99 return numFrames;
|
WaveTableOscillator.h | 34 int getData(long time, float* outBuff, int numFrames, int outChans);
|
/cts/apps/CtsVerifier/src/com/android/cts/verifier/audio/audiolib/ |
AudioFiller.java | 31 * @param numFrames The number of frames of audio to provide. 36 public int fill(float[] buffer, int numFrames, int numChans);
|
WaveScopeView.java | 55 public void setPCM16Buff(short[] smpl16Buff, int numChans, int numFrames) { 60 mNumFrames = numFrames; 67 public void setPCMFloatBuff(float[] smplFloatBuff, int numChans, int numFrames) { 72 mNumFrames = numFrames; 111 * @param numFrames The number of FRAMES in the specified sample block. 116 private void drawChannel16(Canvas cvs, short[] samples, int numFrames, int numChans, 122 for (int frame = 0; frame < numFrames; frame++) { 140 * @param numFrames The number of FRAMES in the specified sample block. 145 private void drawChannelFloat(Canvas cvs, float[] samples, int numFrames, int numChans, 151 for (int frame = 0; frame < numFrames; frame++) [all...] |
/cts/tests/openglperf2/jni/reference/ |
GLReference.cpp | 26 JNIEnv* env, jclass /*clazz*/, jobject assetManager, jobject surface, jint numFrames, 31 if (numFrames > (ReferenceRenderer::FRAMES_PER_SCENE * ReferenceRenderer::NUM_SCENES)) { 42 double updates[numFrames]; 43 double renders[numFrames]; 44 for (int i = 0; i < numFrames && success; i++) { 54 env->SetDoubleArrayRegion(updateTimes, 0, numFrames, updates); 55 env->SetDoubleArrayRegion(renderTimes, 0, numFrames, renders);
|
/frameworks/av/media/libaaudio/src/fifo/ |
FifoControllerBase.cpp | 44 void FifoControllerBase::advanceReadIndex(fifo_frames_t numFrames) { 45 setReadCounter(getReadCounter() + numFrames); 57 void FifoControllerBase::advanceWriteIndex(fifo_frames_t numFrames) { 58 setWriteCounter(getWriteCounter() + numFrames);
|
FifoControllerBase.h | 83 * @param numFrames number of frames to advance the read index 85 void advanceReadIndex(fifo_frames_t numFrames); 98 * @param numFrames number of frames to advance the write index 100 void advanceWriteIndex(fifo_frames_t numFrames);
|
FifoBuffer.cpp | 86 wrappingBuffer->numFrames[1] = 0; 93 wrappingBuffer->numFrames[0] = mFrameCapacity - startIndex; 95 wrappingBuffer->numFrames[1] = mFrameCapacity - startIndex; 99 wrappingBuffer->numFrames[0] = framesAvailable; 103 wrappingBuffer->numFrames[0] = 0; 122 fifo_frames_t FifoBuffer::read(void *buffer, fifo_frames_t numFrames) { 125 fifo_frames_t framesLeft = numFrames; 133 fifo_frames_t framesAvailable = wrappingBuffer.numFrames[partIndex]; 148 fifo_frames_t framesRead = numFrames - framesLeft; 153 fifo_frames_t FifoBuffer::write(const void *buffer, fifo_frames_t numFrames) { [all...] |
/frameworks/av/media/libaaudio/examples/utils/ |
SineGenerator.h | 37 double numFrames = seconds * mFrameRate; 38 mUpScaler = pow((frequencyHigh / frequencyLow), (1.0 / numFrames)); 44 void render(int16_t *buffer, int32_t channelStride, int32_t numFrames) { 46 for (int i = 0; i < numFrames; i++) { 52 void render(float *buffer, int32_t channelStride, int32_t numFrames) { 54 for (int i = 0; i < numFrames; i++) {
|
AAudioSimplePlayer.h | 190 const int numFrames = 32; 191 float zeros[numFrames * samplesPerFrame]; 193 aaudio_result_t result = numFrames; 194 while (result == numFrames) { 195 result = AAudioStream_write(mStream, zeros, numFrames, 0); 259 int32_t numFrames 270 sineData->framesTotal += numFrames; 296 if (numFrames > sineData->maxNumFrames) { 297 sineData->maxNumFrames = numFrames; 299 if (numFrames < sineData->minNumFrames) [all...] |
/frameworks/native/libs/gui/include/gui/ |
OccupancyTracker.h | 44 numFrames(0), 51 numFrames(_numFrames), 60 size_t numFrames; 83 numFrames = 0; 88 size_t numFrames;
|
/frameworks/av/media/libaaudio/src/client/ |
AudioStreamInternalCapture.h | 36 int32_t numFrames, 59 int32_t numFrames, 67 * @param numFrames 70 aaudio_result_t readNowWithConversion(void *buffer, int32_t numFrames);
|
AudioStreamInternalPlay.h | 41 int32_t numFrames, 71 int32_t numFrames, 78 * @param numFrames 82 int32_t numFrames);
|
AudioStreamInternalCapture.cpp | 58 aaudio_result_t AudioStreamInternalCapture::read(void *buffer, int32_t numFrames, 61 return processData(buffer, numFrames, timeoutNanoseconds); 65 aaudio_result_t AudioStreamInternalCapture::processDataNow(void *buffer, int32_t numFrames, 112 //ALOGD("AudioStreamInternalCapture::processDataNow() - readNowWithConversion(%d)", numFrames); 113 int32_t framesProcessed = readNowWithConversion(buffer, numFrames); 115 // numFrames, framesProcessed); 154 int32_t numFrames) { 156 // buffer, numFrames); 159 int32_t framesLeft = numFrames; 166 int32_t framesAvailable = wrappingBuffer.numFrames[partIndex] [all...] |
/frameworks/av/media/libeffects/downmix/ |
EffectDownmix.h | 101 void Downmix_foldFromQuad(LVM_FLOAT *pSrc, LVM_FLOAT *pDst, size_t numFrames, bool accumulate); 102 void Downmix_foldFrom5Point1(LVM_FLOAT *pSrc, LVM_FLOAT *pDst, size_t numFrames, bool accumulate); 103 void Downmix_foldFrom7Point1(LVM_FLOAT *pSrc, LVM_FLOAT *pDst, size_t numFrames, bool accumulate); 105 uint32_t mask, LVM_FLOAT *pSrc, LVM_FLOAT *pDst, size_t numFrames, bool accumulate); 107 void Downmix_foldFromQuad(int16_t *pSrc, int16_t*pDst, size_t numFrames, bool accumulate); 108 void Downmix_foldFrom5Point1(int16_t *pSrc, int16_t*pDst, size_t numFrames, bool accumulate); 109 void Downmix_foldFrom7Point1(int16_t *pSrc, int16_t*pDst, size_t numFrames, bool accumulate); 111 uint32_t mask, int16_t *pSrc, int16_t*pDst, size_t numFrames, bool accumulate);
|
EffectDownmix.c | 336 size_t numFrames = outBuffer->frameCount; 346 while (numFrames) { 351 numFrames--; 354 while (numFrames) { 359 numFrames--; 368 downmixInputChannelMask, pSrc, pDst, numFrames, accumulate)) { 378 Downmix_foldFromQuad(pSrc, pDst, numFrames, accumulate); 382 Downmix_foldFrom5Point1(pSrc, pDst, numFrames, accumulate); 385 Downmix_foldFrom7Point1(pSrc, pDst, numFrames, accumulate); 389 downmixInputChannelMask, pSrc, pDst, numFrames, accumulate)) [all...] |
/external/webrtc/webrtc/modules/video_processing/test/ |
writeYUV420file.m | 9 numFrames=size(Y,3); 11 for k=1:numFrames
|
readYUV420file.m | 21 numFrames = floor(fileLen/frameSizeBytes); 23 Y=uint8(zeros(height,width,numFrames)); 24 U=uint8(zeros(height/2,width/2,numFrames)); 25 V=uint8(zeros(height/2,width/2,numFrames)); 29 for k=1:numFrames
|
/developers/build/prebuilts/gradle/MidiScope/Application/src/main/java/com/example/android/common/midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/developers/build/prebuilts/gradle/MidiSynth/Application/src/main/java/com/example/android/common/midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/developers/samples/android/common/src/java/com/example/android/common/midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/development/samples/browseable/MidiScope/src/com.example.android.common.midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|
/development/samples/browseable/MidiSynth/src/com.example.android.common.midi/synth/ |
SynthVoice.java | 51 int numFrames = outputBuffer.length / samplesPerFrame; 52 for (int i = 0; i < numFrames; i++) {
|