/external/webrtc/src/modules/audio_processing/agc/ |
digital_agc.h | 68 const WebRtc_Word16 *in, // (i) Speech signal
|
/external/webrtc/src/modules/audio_processing/ |
splitting_filter.h | 54 * - out_data : super-wb speech signal
|
/frameworks/av/media/libstagefright/codecs/amrnb/common/include/ |
n_proc.h | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec
|
/frameworks/av/media/libstagefright/codecs/amrnb/dec/src/ |
dtx_dec.cpp | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec 164 dtx_dec.c, UMTS GSM AMR speech codec, R99 - Version 3.3.0, December 12, 2001 357 synth = synthesised speech 382 dtx_dec.c, UMTS GSM AMR speech codec, R99 - Version 3.3.0, December 12, 2001 396 Word16 synth[], // o : synthesised speech 431 // This function is called if synthesis state is not SPEECH 436 // new_state (SPEECH, DTX, DTX_MUTE) 488 // make log_en speech coder mode independent 580 Set_zero(lsfState->past_r_q, M); // reset for next speech frame 596 // or when SID_UPD has been received right after SPEECH [all...] |
bgnscd.cpp | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec 120 bgnscd.c, UMTS GSM AMR speech codec, R99 - Version 3.2.0, March 2, 2001 192 speech[] = synthesis speech frame (Word16) 216 Characterize synthesis speech and detect background noise. 226 bgnscd.c, UMTS GSM AMR speech codec, R99 - Version 3.2.0, March 2, 2001 233 Word16 speech[], // o : synthesis speech frame 254 s = L_mac (s, speech[i], speech[i]) [all...] |
bgnscd.h | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec 106 /* history vector of past synthesis speech energy */ 110 Word16 bgHangover; /* counter; number of frames after last speech frame */ 144 * Purpose : Charaterice synthesis speech and detect background noise 149 Word16 speech[], /* o : synthesis speech frame */
|
/frameworks/av/media/libstagefright/codecs/amrnb/enc/src/ |
cod_amr.cpp | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec 187 cod_amr.c, UMTS GSM AMR speech codec, R99 - Version 3.2.0, March 2, 2001 361 cod_amr.c, UMTS GSM AMR speech codec, R99 - Version 3.2.0, March 2, 2001 376 * Initialize pointers to speech vector. * 379 st->new_speech = st->old_speech + L_TOTAL - L_FRAME; // New speech 381 st->speech = st->new_speech - L_NEXT; // Present frame 474 * Initialize pointers to speech vector. * 477 st->new_speech = st->old_speech + L_TOTAL - L_FRAME; /* New speech */ 479 st->speech = st->new_speech - L_NEXT; /* Present frame */ 577 cod_amr.c, UMTS GSM AMR speech codec, R99 - Version 3.2.0, March 2, 200 [all...] |
pre_big.cpp | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec 100 frameOffset = Word16 -- Start position in speech vector, Q0 101 speech[] = array of type Word16 -- speech, Q0 105 wsp = array of type Word16 -- weighted speech Q0 129 pre_big.c, UMTS GSM AMR speech codec, R99 - Version 3.2.0, March 2, 2001 164 Word16 frameOffset, /* i : Start position in speech vector, Q0 */ 165 Word16 speech[], /* i : speech, Q0 */ 167 Word16 wsp[], /* o : weighted speech Q0 * [all...] |
amrencode.cpp | 22 ANSI-C code for the Adaptive Multi-Rate (AMR) speech codec 494 pEncInput = pointer to the input speech samples (Word16) 526 input speech samples, and returns the newly encoded bit stream in the buffer 542 succeeding octets contain the packed encoded speech bits. The total number of 549 and the most significant nibble contains the first 4 encoded speech bits. The 550 suceeding octets contain the packed encoded speech bits. The total number of 555 encoded speech parameters, then, sid_sync is called to determine the transmit 558 followed by the encoded speech parameters. The codec mode information is 559 stored immediately after the MAX_SERIAL_SIZE encoded speech parameters. If 561 speech parameters, and codec mode are stored in the same order as befor [all...] |
/frameworks/base/core/java/android/speech/tts/ |
FileSynthesisCallback.java | 16 package android.speech.tts; 19 import android.speech.tts.TextToSpeechService.UtteranceProgressDispatcher; 28 * Speech synthesis request that writes the audio to a WAV file. 95 // File will be closed by the SpeechItem in the speech service.
|
SynthesisPlaybackQueueItem.java | 16 package android.speech.tts; 18 import android.speech.tts.TextToSpeechService.AudioOutputParams; 19 import android.speech.tts.TextToSpeechService.UtteranceProgressDispatcher;
|
TextToSpeechService.java | 16 package android.speech.tts; 35 import android.speech.tts.TextToSpeech.Engine; 198 * Notifies the engine that it should load a speech synthesis language. There is no guarantee 220 * Notifies the service that it should stop any in-progress speech synthesis. 221 * This method can be called even if no speech synthesis is currently in progress. 228 * Tells the service to synthesize speech from the given text. This method 230 * clients ({@link android.speech.tts.TextToSpeech}). Called on the synthesis 347 * Notifies the engine that it should load a speech synthesis voice. There is no guarantee 516 // Don't process any more speech items 518 // Stop the current speech ite [all...] |
TextToSpeech.java | 16 package android.speech.tts; 52 * Synthesizes speech from text for immediate playback or to create a sound file. 163 * of all the text in the speech queue. 172 "android.speech.tts.TTS_QUEUE_PROCESSING_COMPLETED"; 205 * Constants and parameter names for controlling text-to-speech. These include: 228 * Default speech rate. 271 * Default audio stream used when playing synthesized speech. 325 * Name under which a text to speech engine publishes information about itself. 330 public static final String SERVICE_META_DATA = "android.speech.tts"; 344 "android.speech.tts.engine.INSTALL_TTS_DATA" [all...] |
/frameworks/base/docs/html/training/accessibility/ |
index.jd | 51 use a text-to-speech engine to speak to the user.</dd>
|
/packages/apps/Messaging/src/com/android/messaging/ui/mediapicker/ |
AudioLevelSource.java | 23 * Keeps track of the speech level as last observed by the recognition
|
/packages/apps/Settings/src/com/android/settings/ |
VoiceInputOutputSettings.java | 22 import android.speech.tts.TtsEngines;
|
/packages/inputmethods/OpenWnn/src/jp/co/omronsoft/openwnn/ |
WnnClause.java | 31 * @param posTag The part of speech of the clause
|
/frameworks/base/docs/html/about/versions/ |
android-1.5-highlights.jd | 189 <li>Speech recognition framework 191 <li>Support for using speech recognition libraries via Intent. See {@link 192 android.speech.RecognizerIntent RecognizerIntent}.</li>
|
android-2.3.3.jd | 165 <h3 id="speech">Speech recognition</h3> 167 <p>The speech-recognition API includes new constants to let you manage voice 169 use of speech recognition, you could use them to offer a different view of voice 171 android.speech.RecognizerResultsIntent}.</p>
|
/frameworks/av/media/libstagefright/codecs/amrwb/src/ |
pvamrwbdecoder.cpp | 22 ANSI-C code for the Adaptive Multi-Rate - Wideband (AMR-WB) speech codec 49 int16 synth16k[], output: synthesis speech 59 speech frames for wideband speech signals. 284 int16 synth16k[], /* output: synthesis speech */ 345 /* find the new DTX state SPEECH OR DTX */ 349 if (newDTXState != SPEECH) 353 /* SPEECH action state machine */ 389 /* If this frame is the first speech frame after CNI period, 391 * on whether there was DTX muting before start of speech or no [all...] |
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/ |
voAMRWBEnc.c | 23 * wideband speech signals. * 153 Word16 speech16k[], /* input : 320 new speech samples (at 16 kHz) */ 162 /* Speech vector */ 164 Word16 *new_speech, *speech, *p_window; local 166 /* Weighted speech vector */ 175 Word16 r_h[M + 1], r_l[M + 1]; /* Autocorrelations of windowed speech */ 221 * Initialize pointers to speech vector. * 226 * <------- Total speech buffer (L_TOTAL) ------> * 230 * p_window | <----- new speech (L_FRAME) ----> * 232 * speech | * [all...] |
/frameworks/support/v17/leanback/src/android/support/v17/leanback/widget/ |
SearchBar.java | 27 import android.speech.RecognitionListener; 28 import android.speech.RecognizerIntent; 29 import android.speech.SpeechRecognizer; 211 // don't propagate event during speech recognition. 421 * Sets the speech recognizer to be used when doing voice search. The Activity/Fragment is in 440 throw new IllegalStateException("Can't have speech recognizer and request"); 445 * Sets the speech recognition callback. 450 throw new IllegalStateException("Can't have speech recognizer and request"); 503 * Stops the speech recognition, if already started. 621 Log.w(TAG, "recognizer speech timeout") [all...] |
/external/speex/include/speex/ |
speex.h | 307 /** Uses an existing encoder state to encode one frame of speech pointed to by 318 /** Uses an existing encoder state to encode one frame of speech pointed to by 353 /** Uses an existing decoder state to decode one frame of speech from 354 * bit-stream bits. The output speech is saved written to out. 363 /** Uses an existing decoder state to decode one frame of speech from 364 * bit-stream bits. The output speech is saved written to out.
|
speex_preprocess.h | 75 * @return Bool value for voice activity (1 for speech, 0 for noise/silence), ONLY if VAD turned on. 205 /* Can't set speech probability */ 206 /** Get speech probability in last frame (int32). */
|
/external/libopus/silk/fixed/ |
main_FIX.h | 97 const opus_int16 x[] /* I Speech signal */ 137 const opus_int16 x[], /* I Speech signal */ 146 const opus_int16 x[], /* I Speech signal */
|