Method for processing the output of a speech recognizer
원문보기
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G10L-015/18
G06F-017/27
G10L-015/22
H04L-012/28
출원번호
US-0921780
(2015-10-23)
등록번호
US-9753912
(2017-09-05)
발명자
/ 주소
Roy, Philippe
Lagassey, Paul J.
출원인 / 주소
Great Northern Research, LLC
대리인 / 주소
Hoffberg, Steven M.
인용정보
피인용 횟수 :
0인용 특허 :
199
초록▼
A method for processing speech, comprising semantically parsing a received natural language speech input with respect to a plurality of predetermined command grammars in an automated speech processing system; determining if the parsed speech input unambiguously corresponds to a command and is suffic
A method for processing speech, comprising semantically parsing a received natural language speech input with respect to a plurality of predetermined command grammars in an automated speech processing system; determining if the parsed speech input unambiguously corresponds to a command and is sufficiently complete for reliable processing, then processing the command; if the speech input ambiguously corresponds to a single command or is not sufficiently complete for reliable processing, then prompting a user for further speech input to reduce ambiguity or increase completeness, in dependence on a relationship of previously received speech input and at least one command grammar of the plurality of predetermined command grammars, reparsing the further speech input in conjunction with previously parsed speech input, and iterating as necessary. The system also monitors abort, fail or cancel conditions in the speech input.
대표청구항▼
1. A method for processing speech, comprising: receiving information representing speech input;parsing the received information representing speech input with respect to a plurality of predetermined command contexts, the plurality of predetermined command contexts comprising command contexts of a pl
1. A method for processing speech, comprising: receiving information representing speech input;parsing the received information representing speech input with respect to a plurality of predetermined command contexts, the plurality of predetermined command contexts comprising command contexts of a plurality of different applications, in an automated speech information processing system;determining, with at least one automated processor: if the parsed received information representing speech input corresponds to a command of a respective predetermined command context for execution of the command, and based on the determination that the parsed received information representing speech input corresponds to a command of the respective predetermined command context for reliable execution, then executing the command by a respective application associated with the respective predetermined command context;if the parsed received information representing speech input cannot be executed, and based on the determination that the parsed received information representing speech input cannot be executed, then prompting for further information representing speech input, to reduce ambiguity or increase completeness, in dependence on previously received information representing speech input and the plurality of predetermined command contexts; andif an abort, fail or cancel condition is detected in further information representing speech input, and based on the determination that the abort, fail or cancel condition is detected in the further information representing speech input, then exiting said determining. 2. The method according to claim 1, wherein the information representing speech input comprises information derived from a natural language speech input from a human. 3. The method according to claim 1, wherein the plurality of predetermined command contexts represent a plurality of modes of the plurality of respective different applications. 4. The method according to claim 1, wherein the determining if the parsed received information representing speech input cannot be executed, comprises determining an inconsistency of the received information representing speech input with respect to the plurality of predetermined command contexts. 5. The method according to claim 1, wherein the plurality of predetermined command contexts comprises a home automation task. 6. The method according to claim 1, wherein the plurality of predetermined command contexts comprises at least one task associated with a future time. 7. The method according to claim 1, wherein each of the plurality of predetermined command contexts has at least one entry in an associated command dictionary. 8. The method according to claim 7, wherein each entry in the associated command dictionary has an associated command grammar. 9. The method according to claim 8, wherein the determining if the parsed received information representing speech input corresponds to a command of the respective predetermined command context for execution, comprises associating parsed portions of the received information representing speech input with the respective command grammars. 10. The method according to claim 1, wherein the at least one automated processor processes the parsed received information representing speech input according to a statistical process. 11. A method for processing speech, comprising: receiving speech input;automatically processing the received speech input with at least one automated processor, to parse the speech input, in accordance with a plurality of available command contexts, associated with a plurality of different applications which accept respective commands, and if a correspondence of previously received speech input to any single command is incomplete for execution by any of the plurality of different applications, automatically prompting a human user for further speech input, with a prompt adapted to solicit information from the human user to reduce ambiguity or increase completeness with respect to respective commands accepted by the available command contexts;automatically disregarding previously receive speech input if an abort, fail or cancel condition is detected by the at least one automated processor in further speech input; andbased on a result of the automatically processing, selectively processing the command with the at least one automated command processor. 12. The method according to claim 11, wherein the automatically processing the received speech input comprises a statistical speech recognition process, performed on natural language human speech. 13. The method according to claim 11, wherein each of the available command contexts has at least one entry in an associated command dictionary, and each entry in the associated command dictionary has an associated command grammar, andthe determining if the correspondence of the previously received speech input to any single command is incomplete for execution, comprises associating portions of the parsed speech input with the respective command grammars. 14. The method according to claim 12, wherein the speech input comprises a natural language speech input from a human. 15. The method according to claim 12, wherein the plurality of predetermined command contexts represent a plurality of modes of the plurality of different applications. 16. The method according to claim 12, wherein the determining if the parsed speech input does not represent a command that can be executed, comprises determining an inconsistency of the received speech input with respect to the plurality of predetermined command contexts. 17. The method according to claim 12, wherein the plurality of predetermined command contexts comprises a home automation task. 18. The method according to claim 17, wherein each entry in the associated command dictionary has an associated command grammar. 19. The method according to claim 12, wherein the plurality of predetermined command contexts comprises at least one task associated with a future time. 20. The method according to claim 19, wherein the determining if the parsed speech input represents a command of the respective predetermined command context that can be executed, comprises associating parsed portions of the speech input with the respective command grammars. 21. The method according to claim 12, wherein each of the plurality of predetermined command contexts has at least one entry in an associated command dictionary. 22. The method according to claim 12, wherein the automatically determining is performed using at least one automated speech processor, which processes the parsed speech input according to a statistical process. 23. A speech processing method, comprising: receiving a speech input;automatically parsing the speech input, in an automated speech information processing system, with respect to a plurality of predetermined command contexts associated with a plurality of different applications which each accept commands for execution;automatically determining: if the parsed speech input represents a command of a respective predetermined command context that can be executed,then executing the command by the respective application which is associated with the respective predetermined command context,else, automatically determining: if the parsed received speech input does not represent a command that can be executed by any respective application which is associated with the plurality of predetermined command contexts,then prompting for further information representing speech input, to reduce ambiguity or increase completeness, in dependence on previously received speech input and the plurality of predetermined command contexts, andautomatically determining: if an abort, fail or cancel condition is detected in further speech input,then exiting said determining.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (199)
Julia, Luc; Voutsas, Dimitris; Cheyer, Adam, Accessing network-based electronic information through scripted online interfaces using spoken input.
Van Vliembergen Eduardus J. W. (Venlo NLX) Heemels Robertus P. E. H. (Roermond NLX) Cremers Louis M. G. (Venlo NLX) Hommersom Frederik J. (Venlo NLX) Gerritsen Jan (Nijmegen NLX), Apparatus and method for syntactic signal analysis.
Deligne Sabine ; Sagisaka Yoshinori,JPX ; Nakajima Hideharu,JPX, Apparatus for generating a statistical sequence model called class bi-multigram model with bigram dependencies assumed between adjacent sequences.
Kamatani, Satoshi; Chino, Tetsuro; Furihata, Kentaro, Apparatus, system, method, and computer program product for resolving ambiguities in translations.
Shieber Stuart M. ; Armstrong John ; Baptista Rafael Jose ; Bentz Bryan A. ; Ganong ; III William F. ; Selesky Donald Bryant, Command parsing and rewrite system.
Kamatani, Satoshi; Chino, Tetsuo, Communication support apparatus and computer program product for supporting communication by performing translation between languages.
Kamatani, Satoshi; Chino, Tetsuro; Kuroda, Yuka, Communication support apparatus and computer program product for supporting communication by performing translation between languages.
Steel,Samuel W. D.; Kruschwitz,Udo; Webb,Nicholas J.; De Roeck,Anne N.; Scott,Paul D.; Turner,Raymond; Tsui,Kwok C.; Wobcke,Wayne R.; Azvine,Behnam, Index to a semi-structured database.
Bennett, Ian M.; Babu, Bandi Ramesh; Morkhandikar, Kishor; Gururaj, Pallaki, Interactive speech based learning/training system formulating search queries based on natural language parsing of recognized user queries.
Matheson,Caroline, Man-machine dialogue system, controls dialogue between system and user using dialogue specification employing augmented transition networks propagating tokens.
Colier Ronald L. (T/L Apartments ; 16 Cheshire Dr. ; Apt. 121 Pittsfield MA 01201), Method and apparatus adapted for an audibly-driven, handheld, keyless and mouseless computer for performing a user-cente.
Monaco Peter C. ; Ehrlich Steven C. ; Ghosh Debajit ; Klenk Mark ; Sinai Julian ; Thirumalai Madhavan ; Gupta Sundeep, Method and apparatus for creating modifiable and combinable speech objects for acquiring information from a speaker in an interactive voice response system.
Tal Peter (53 Driftwood Dr. Port Washington NY 11050), Method and apparatus for uniquely identifying individuals by particular physical characteristics and security system uti.
Fischer,Uwe; Hoffmann,Stefan; Kriechbaum,Werner; Stenzel,Gerhard, Method and system for the automatic generation of multi-lingual synchronized sub-titles for audiovisual data.
Michael S. Phillips ; Mark A. Fanty ; Krishna K. Govindarajan, Method and system of reviewing the behavior of an interactive speech recognition application.
Noyes Dallas B. (2500 George Washington Way ; #124 Richland WA 99352), Method for representation of knowledge in a computer as a network database system.
Battle James Thomas ; Hung Andy C. ; Purcell Stephen C., Multimedia processor using variable length instructions with opcode specification of source operand as result of prior i.
Loatman Robert B. (Vienna VA) Post Stephen D. (McLean VA) Yang Chih-King (Rockville MD) Hermansen John C. (Catharpin VA), Natural language understanding system.
Janek Gabor,HUX ; Wutte Heribert,AUX ; Grabherr Manfred, Product including a speech recognition device and method of generating a command lexicon for a speech recognition device.
Ostendorf Mari ; Singer Harald,JPX, Speaker-independent model generation apparatus and speech recognition apparatus each equipped with means for splitting.
Morgan Scott Anthony ; Roberts David John,GBX ; Swearingen Craig Ardner ; Tannenbaum Alan Richard, Speech command input recognition system for interactive computer display with term weighting means used in interpreting potential commands from relevant speech terms.
Gagnon, Jean; Roy, Philippe; Lagassey, Paul J., Speech interface system and method for control and interaction with applications on a computing system.
Ohmori,Kumiko; Higashida,Masanobu; Mizusawa,Noriko, Speech recognition based interactive information retrieval scheme using dialogue control to reduce user stress.
Tang Donald T.,CNX ; Shen Li Qin,CNX ; Zhu Xiao Jin,CNX, Speech recognition method and system for recognizing single or un-correlated Chinese characters.
Marx Matthew T. ; Carter Jerry K. ; Phillips Michael S. ; Holthouse Mark A. ; Seabury Stephen D. ; Elizondo-Cecenas Jose L. ; Phaneuf Brett D., System and method for developing interactive speech applications.
Dantzig,Paul M.; Filepp,Robert; Liu,Yew Huey, System and method for generating and presenting multi-modal applications from intent-based markup scripts.
Maes, Stephane Herman; Neti, Chalapathy Venkata, System and method for multi-modal focus detection, referential ambiguity resolution and mood classification using multi-modal input.
Coffman, Daniel M.; Hosn, Rafah A.; Kleindienst, Jan; Maes, Stephane H.; Raman, Thiruvilwamalai V., System and method for providing dialog management and arbitration in a multi-modal environment.
Eberle, Hannes; Leon, Christopher S.; Maass, Bodo; Patnaik, Anurag; Santa Ana, Alberto; Zirngibl, Michael, System and method for the creation and automatic deployment of personalized, dynamic and interactive inbound and outbound voice services, with real-time interactive voice database queries.
Maes, Stephane H.; Lubensky, David M.; Sakrajda, Andrzej, Universal IP-based and scalable architectures across conversational applications using web services for speech and audio processing resources.
Ackley H. Sprague ; Maltsev Pavel A. ; Ohanian Michael, Universal data input and processing device, such as universal point-of-sale device for inputting and processing bar code symbols, document images, and other data.
Ball, Thomas J.; Cox, Kenneth Charles; Grinter, Rebecca Elizabeth; Hibino, Stacie Lynn; Jagadeesan, Lalita Jategaonkar; Mantilla, David Alejandro, User interface for translating natural language inquiries into database queries and data presentations.
Brant Arthur ; Mandell Kenneth ; Rader R. Scott ; Walsh Alexander ; deJuan ; Jr. Eugene ; Greenberg Robert, Voice command and control medical care system.
Alpdemir, Ahmet; James, Arthur, Voice-interactive marketplace providing promotion and promotion tracking, loyalty reward and redemption, and other features.
Lim, Kang S.; Nguyen, Joseph A., Voice-recognition-based methods for establishing outbound communication through a unified messaging system including intelligent calendar interface.
Mori,Shinsuke; Nishimura,Masafumi; Itoh,Nobuyasu, Word predicting method, voice recognition method, and voice recognition apparatus and program using the same methods.
Kaplan, Boris, Working method for treatment of abstract objects (the thought-substances) of the system of artificial intelligence of a cyborg or an android for the pointer-oriented object acquisition method for abstract treatment of information of this system based on a natural language.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.