/external/chromium_org/chrome/browser/resources/chromeos/chromevox/host/chrome/ |
tts_background.js | 23 * @param {Object} properties Speech properties to use for this utterance. 25 cvox.Utterance = function(textString, properties) { 28 this.id = cvox.Utterance.nextUtteranceId_++; 32 * The next utterance id to use. 36 cvox.Utterance.nextUtteranceId_ = 1; 155 * The current utterance. 156 * @type {cvox.Utterance} 162 * The utterance queue. 163 * @type {Array.<cvox.Utterance>} 298 var utterance = new cvox.Utterance(textString, mergedProperties) [all...] |
tts.js | 112 * @param {Object=} properties Speech properties to use for this utterance.
|
/external/chromium_org/chrome/browser/speech/ |
tts_chromeos.cc | 26 const std::string& utterance,
|
tts_android.cc | 42 const std::string& utterance, 50 base::android::ConvertUTF8ToJavaString(env, utterance).Release(), 56 utterance_ = utterance;
|
tts_android.h | 18 const std::string& utterance,
|
tts_controller.h | 18 class Utterance; 46 // Returns true if this event type is one that indicates an utterance 50 // The continuous parameters that apply to a given utterance. 84 virtual void OnTtsEvent(Utterance* utterance, 98 // One speech utterance. 99 class Utterance { 101 // Construct an utterance given a profile and a completion task to call 102 // when the utterance is done speaking. Before speaking this utterance, [all...] |
tts_controller_unittest.cc | 23 const std::string& utterance, 52 Utterance* utterance1 = new Utterance(NULL); 57 Utterance* utterance2 = new Utterance(NULL);
|
tts_linux.cc | 39 const std::string& utterance, 79 // These apply to the current utterance only. 167 const std::string& utterance, 195 utterance_ = utterance; 198 if (libspeechd_loader_.spd_say(conn_, SPD_TEXT, utterance.c_str()) == -1) {
|
tts_win.cc | 24 const std::string& utterance, 52 // These apply to the current utterance only. 126 // utterance are ignored.
|
/external/chromium_org/chrome/renderer/ |
tts_dispatcher.h | 29 // the utterance id (which is globally unique) matches. 44 virtual void speak(const blink::WebSpeechSynthesisUtterance& utterance) 69 // Next utterance id, used to map response IPCs to utterance objects. 72 // Map from id to utterance objects.
|
/external/chromium_org/chrome/common/extensions/docs/examples/extensions/talking_alarm_clock/ |
common.js | 144 var utterance = phraseTemplate.replace(/\$TIME/g, timeString); 145 speak(utterance);
|
/external/chromium_org/third_party/WebKit/Source/modules/speech/testing/ |
PlatformSpeechSynthesizerMock.cpp | 77 void PlatformSpeechSynthesizerMock::speak(PlatformSpeechSynthesisUtterance* utterance) 80 m_utterance = utterance;
|
/external/chromium_org/chrome/browser/chromeos/accessibility/ |
spoken_feedback_browsertest.cc | 111 std::string utterance = monitor.GetNextUtterance(); local 112 VLOG(0) << "Got utterance: " << utterance; 113 if (utterance == "Bookmarks,") 189 std::string utterance = monitor.GetNextUtterance(); local 190 VLOG(0) << "Got utterance: " << utterance; 191 if (utterance == "text box")
|
/external/chromium_org/chrome/browser/speech/extension_api/ |
tts_extension_apitest.cc | 57 const std::string& utterance, 100 const std::string& utterance, 108 false, utterance_id, TTS_EVENT_END, utterance.size(), 115 const std::string& utterance, 123 true, utterance_id, TTS_EVENT_END, utterance.size(), std::string()), 128 const std::string& utterance, 132 for (int i = 0; i < static_cast<int>(utterance.size()); i++) { 133 if (i == 0 || utterance[i - 1] == ' ') { 239 // One utterance starts speaking, and then a second interrupts. 245 // Expect the second utterance and allow it to finish [all...] |
/external/chromium_org/content/common/ |
speech_recognition_messages.h | 29 IPC_STRUCT_TRAITS_MEMBER(utterance)
|
/external/chromium_org/chrome/browser/resources/chromeos/chromevox/chromevox/background/ |
accessibility_api_handler.js | 262 // Queue the next utterance because a window opening is always followed 384 this.tts.speak(description.utterance, 406 this.tts.speak(description.utterance, 490 this.tts.speak(description.utterance, 528 var utterance = this.idleSpeechQueue_.shift(); 530 this.tts.speak(utterance, 537 * utterance to speak, text to braille, and an earcon to play to describe it. 541 * @return {Object} An object containing a string field |utterance|, object 692 description.utterance = s; 727 var utterance = '' [all...] |
/external/chromium_org/content/browser/speech/ |
google_one_shot_remote_engine.cc | 32 const char* const kUtteranceString = "utterance"; 125 base::string16 utterance; local 127 if (!hypothesis_value->GetString(kUtteranceString, &utterance)) { 128 LOG(WARNING) << "ParseServerResponse: Missing utterance value."; 135 result->hypotheses.push_back(SpeechRecognitionHypothesis(utterance,
|
google_streaming_remote_engine_unittest.cc | 428 proto_alternative->set_transcript(base::UTF16ToUTF8(hypothesis.utterance)); 476 if (hyp_a.utterance != hyp_b.utterance ||
|
/external/srec/srec/Recognizer/src/ |
RecognizerImpl.c | 537 /* Create an utterance object */ 538 impl->utterance = CA_AllocateUtterance(); 539 if (impl->utterance == NULL) 545 CA_InitUtteranceForFrontend(impl->utterance, frontendParams); 546 CA_AttachCMStoUtterance(impl->wavein, impl->utterance); 563 if (impl->utterance != NULL) 565 CA_ClearUtterance(impl->utterance); 566 CA_FreeUtterance(impl->utterance); 567 impl->utterance = NULL; 671 impl->utterance = NULL [all...] |
/external/chromium_org/content/test/ |
mock_google_streaming_server.cc | 96 proto_alternative->set_transcript(base::UTF16ToUTF8(hypothesis.utterance));
|
/external/chromium_org/chrome/browser/resources/chromeos/chromevox/common/ |
editable_text.js | 649 * to the end, if any. It should be spoken at the end of the utterance 661 var utterance = ''; 665 utterance = inserted; 678 utterance = evt.value.substr(index, prefixLen + 1 - index); 680 utterance = inserted; 687 utterance = cvox.ChromeVoxEditableTextBase.eventTypingEcho ? '' : 691 utterance = deleted + ', deleted'; 693 utterance = deleted; 696 if (autocompleteSuffix && utterance) { 697 utterance += ', ' + autocompleteSuffix [all...] |
/external/chromium_org/chrome/common/extensions/docs/examples/extensions/ttsdebug/ |
ttsdebug.js | 32 function logSpeakCall(utterance, options, callback) { 40 utterance + '\', ' + 43 chrome.tts.speak(utterance, options, callback); 45 chrome.tts.speak(utterance, options); 172 'the baseline utterance, to make sure that speech after ' +
|
/external/chromium_org/chrome/browser/resources/chromeos/chromevox/chromevox/injected/ |
api_implementation.js | 115 * @param {Object} properties Speech properties to use for this utterance. 137 * @param {Object=} properties Speech properties to use for this utterance. 155 * @param {Object=} properties Speech properties to use for this utterance. 172 * @param {Object=} properties Speech properties to use for this utterance.
|
api.js | 120 * @param {Object=} properties Speech properties to use for this utterance. 210 * @param {Object=} properties Speech properties to use for this utterance. 233 * @param {Object=} properties Speech properties to use for this utterance.
|
/external/chromium_org/content/renderer/ |
speech_recognition_dispatcher.cc | 204 transcripts[i] = result.hypotheses[i].utterance;
|