/external/webrtc/webrtc/modules/audio_conference_mixer/source/ |
audio_frame_manipulator.h | 15 class AudioFrame; 17 // Updates the audioFrame's energy (based on its samples). 18 void CalculateEnergy(AudioFrame& audioFrame); 20 // Apply linear step function that ramps in/out the audio samples in audioFrame 21 void RampIn(AudioFrame& audioFrame); 22 void RampOut(AudioFrame& audioFrame);
|
/external/webrtc/webrtc/modules/utility/include/ |
audio_frame_operations.h | 18 class AudioFrame; 33 static int MonoToStereo(AudioFrame* frame); 42 static int StereoToMono(AudioFrame* frame); 46 static void SwapStereoChannels(AudioFrame* frame); 49 static void Mute(AudioFrame& frame); 51 static int Scale(float left, float right, AudioFrame& frame); 53 static int ScaleWithSat(float scale, AudioFrame& frame);
|
/external/webrtc/webrtc/tools/agc/ |
test_utils.h | 15 class AudioFrame; 19 void ApplyGainLinear(float gain, float last_gain, AudioFrame* frame); 20 void ApplyGain(float gain_db, float last_gain_db, AudioFrame* frame); 22 AudioFrame* frame); 24 AudioFrame* frame);
|
fake_agc.h | 25 virtual int Process(const AudioFrame& audio_frame) {
|
test_utils.cc | 29 void ApplyGainLinear(float gain, float last_gain, AudioFrame* frame) { 43 void ApplyGain(float gain_db, float last_gain_db, AudioFrame* frame) { 48 AudioFrame* frame) { 57 AudioFrame* frame) {
|
/external/webrtc/webrtc/modules/audio_coding/acm2/ |
call_statistics.cc | 19 void CallStatistics::DecodedByNetEq(AudioFrame::SpeechType speech_type) { 22 case AudioFrame::kNormalSpeech: { 26 case AudioFrame::kPLC: { 30 case AudioFrame::kCNG: { 34 case AudioFrame::kPLCCNG: { 38 case AudioFrame::kUndefined: {
|
call_statistics.h | 41 void DecodedByNetEq(AudioFrame::SpeechType speech_type);
|
call_statistics_unittest.cc | 36 call_stats.DecodedByNetEq(AudioFrame::kNormalSpeech); 37 call_stats.DecodedByNetEq(AudioFrame::kPLC); 38 call_stats.DecodedByNetEq(AudioFrame::kPLCCNG); 39 call_stats.DecodedByNetEq(AudioFrame::kCNG);
|
acm_receiver.cc | 42 AudioFrame* audio_frame) { 46 audio_frame->vad_activity_ = AudioFrame::kVadActive; 47 audio_frame->speech_type_ = AudioFrame::kNormalSpeech; 51 audio_frame->vad_activity_ = AudioFrame::kVadPassive; 52 audio_frame->speech_type_ = AudioFrame::kNormalSpeech; 56 audio_frame->vad_activity_ = AudioFrame::kVadPassive; 57 audio_frame->speech_type_ = AudioFrame::kCNG; 63 audio_frame->speech_type_ = AudioFrame::kPLC; 67 audio_frame->vad_activity_ = AudioFrame::kVadPassive; 68 audio_frame->speech_type_ = AudioFrame::kPLCCNG [all...] |
/external/webrtc/webrtc/voice_engine/ |
utility.h | 23 class AudioFrame; 31 // This version has an AudioFrame |src_frame| as input and sets the output 34 void RemixAndResample(const AudioFrame& src_frame, 36 AudioFrame* dst_frame); 46 AudioFrame* dst_frame);
|
level_indicator.h | 19 class AudioFrame; 38 void ComputeLevel(const AudioFrame& audioFrame);
|
utility_unittest.cc | 40 AudioFrame src_frame_; 41 AudioFrame dst_frame_; 42 AudioFrame golden_frame_; 48 void SetMonoFrame(AudioFrame* frame, float data, int sample_rate_hz) { 59 void SetMonoFrame(AudioFrame* frame, float data) { 65 void SetStereoFrame(AudioFrame* frame, float left, float right, 78 void SetStereoFrame(AudioFrame* frame, float left, float right) { 82 void VerifyParams(const AudioFrame& ref_frame, const AudioFrame& test_frame) { 91 float ComputeSNR(const AudioFrame& ref_frame, const AudioFrame& test_frame [all...] |
output_mixer.h | 67 AudioFrame* audioFrame); 91 const AudioFrame& generalAudioFrame, 92 const AudioFrame** uniqueAudioFrames, 116 AudioFrame _audioFrame;
|
/external/webrtc/webrtc/modules/audio_conference_mixer/include/ |
audio_conference_mixer_defines.h | 24 // The implementation of this function should update audioFrame with new 29 AudioFrame* audioFrame) = 0; 34 // This function specifies the sampling frequency needed for the AudioFrame 48 // Note that uniqueAudioFrames is an array of AudioFrame pointers with the 51 const AudioFrame& generalAudioFrame, 52 const AudioFrame** uniqueAudioFrames,
|
/external/webrtc/talk/media/base/ |
audioframe.h | 33 class AudioFrame { 35 AudioFrame() 41 AudioFrame(int16_t* audio, size_t audio_length, int sample_freq, bool stereo)
|
/external/webrtc/webrtc/modules/utility/source/ |
coder.h | 20 class AudioFrame; 32 int32_t Decode(AudioFrame& decodedAudio, uint32_t sampFreqHz, 35 int32_t PlayoutData(AudioFrame& decodedAudio, uint16_t& sampFreqHz); 37 int32_t Encode(const AudioFrame& audio, int8_t* encodedData,
|
audio_frame_operations.cc | 25 int AudioFrameOperations::MonoToStereo(AudioFrame* frame) { 29 if ((frame->samples_per_channel_ * 2) >= AudioFrame::kMaxDataSizeSamples) { 34 int16_t data_copy[AudioFrame::kMaxDataSizeSamples]; 51 int AudioFrameOperations::StereoToMono(AudioFrame* frame) { 62 void AudioFrameOperations::SwapStereoChannels(AudioFrame* frame) { 72 void AudioFrameOperations::Mute(AudioFrame& frame) { 77 int AudioFrameOperations::Scale(float left, float right, AudioFrame& frame) { 91 int AudioFrameOperations::ScaleWithSat(float scale, AudioFrame& frame) {
|
audio_frame_operations_unittest.cc | 27 AudioFrame frame_; 30 void SetFrameData(AudioFrame* frame, int16_t left, int16_t right) { 37 void SetFrameData(AudioFrame* frame, int16_t data) { 43 void VerifyFramesAreEqual(const AudioFrame& frame1, const AudioFrame& frame2) { 57 frame_.samples_per_channel_ = AudioFrame::kMaxDataSizeSamples; 65 AudioFrame temp_frame; 69 AudioFrame stereo_frame; 90 AudioFrame temp_frame; 94 AudioFrame mono_frame [all...] |
coder.cc | 51 int32_t AudioCoder::Decode(AudioFrame& decodedAudio, 71 int32_t AudioCoder::PlayoutData(AudioFrame& decodedAudio, 77 int32_t AudioCoder::Encode(const AudioFrame& audio, 83 AudioFrame audioFrame; 84 audioFrame.CopyFrom(audio); 85 audioFrame.timestamp_ = _encodeTimestamp; 86 _encodeTimestamp += static_cast<uint32_t>(audioFrame.samples_per_channel_); 91 if(_acm->Add10MsData((AudioFrame&)audioFrame) == -1 [all...] |
/external/webrtc/webrtc/voice_engine/include/ |
voe_external_media.h | 18 class AudioFrame; 73 AudioFrame* frame) = 0;
|
/external/webrtc/webrtc/modules/audio_processing/vad/ |
pitch_based_vad.h | 21 class AudioFrame;
|
standalone_vad.h | 21 class AudioFrame;
|
/external/webrtc/webrtc/modules/audio_conference_mixer/test/ |
audio_conference_mixer_unittest.cc | 26 const AudioFrame& general_audio_frame, 27 const AudioFrame** unique_audio_frames, 38 int32_t(const int32_t id, AudioFrame* audio_frame)); 40 AudioFrame* fake_frame() { return &fake_frame_; } 43 AudioFrame fake_frame_; 44 int32_t FakeAudioFrame(const int32_t id, AudioFrame* audio_frame) { 122 participants[i].fake_frame()->speech_type_ = AudioFrame::kNormalSpeech; 123 participants[i].fake_frame()->vad_activity_ = AudioFrame::kVadActive; 143 AudioFrame::kVadPassive;
|
/external/webrtc/webrtc/modules/audio_coding/test/ |
PCMFile.h | 38 int32_t Read10MsData(AudioFrame& audio_frame); 41 void Write10MsData(AudioFrame& audio_frame);
|
/external/webrtc/webrtc/modules/audio_processing/agc/ |
agc.h | 20 class AudioFrame;
|