IPC분류정보
국가/구분 |
United States(US) Patent
등록
|
국제특허분류(IPC7판) |
|
출원번호 |
US-0046026
(2002-01-11)
|
등록번호 |
US-7263489
(2007-08-28)
|
발명자
/ 주소 |
- Cohen,Michael H.
- Heck,Larry P.
- Balogh,Jennifer E.
- Riseman,James M.
- Mirghafori,Naghmeh N.
|
출원인 / 주소 |
- Nuance Communications, Inc.
|
대리인 / 주소 |
Blakely Sokoloff Taylor & Zafman LLP
|
인용정보 |
피인용 횟수 :
31 인용 특허 :
37 |
초록
▼
A system which uses automatic speech recognition to provide dialogs with human speakers automatically detects one or more characteristics, which may be characteristics of a speaker, his speech, his environment, or the speech channel used to communicate with the speaker. The characteristic may be det
A system which uses automatic speech recognition to provide dialogs with human speakers automatically detects one or more characteristics, which may be characteristics of a speaker, his speech, his environment, or the speech channel used to communicate with the speaker. The characteristic may be detected either during the dialog or at a later time based on stored data representing the dialog. If the characteristic is detected during the dialog, the dialog can be customized for the speaker at an application level, based on the detected characteristic. The customization may include customization of operations and features such as call routing, error recovery, call flow, content selection, system prompts, or system persona. Data indicative of detected characteristics can be stored and accumulated for many speakers and/or dialogs and analyzed offline to generate a demographic or other type of analysis of the speakers or dialogs with respect to one or more detected characteristics.
대표청구항
▼
What is claimed is: 1. A method comprising: establishing a speech-based dialog between a person and a machine during a call, wherein the person uses a communication device to speak to the machine via a communication channel during the call; automatically detecting a characteristic during the dialog
What is claimed is: 1. A method comprising: establishing a speech-based dialog between a person and a machine during a call, wherein the person uses a communication device to speak to the machine via a communication channel during the call; automatically detecting a characteristic during the dialog in real time, wherein the characteristic is a characteristic of the person, the communication device, the communication channel, or an environment in which the person is located during the dialog, but the characteristic does not uniquely identify the person, the communication device, or any user account; and selecting a destination to which the call should be routed, based on the detected characteristic, and not based on the meaning of any speech or the failure to recognize any speech during the dialog. 2. A method as recited in claim 1, wherein the characteristic is an approximate age of the person. 3. A method as recited in claim 1, wherein the characteristic is the gender of the person. 4. A method as recited in claim 1, wherein the characteristic is a type of speech being spoken by the person. 5. A method as recited in claim 1, wherein the characteristic is an emotional state of the person. 6. A method as recited in claim 1, wherein the characteristic is indicative of the truthfulness of speech of the person. 7. A method as recited in claim 1, wherein the characteristic is an acoustic characteristic. 8. A method as recited in claim 1, wherein the characteristic is indicative of a speech level of the dialog. 9. A method as recited in claim 1, wherein the characteristic is descriptive of a reason the person is experiencing an error. 10. A method as recited in claim 1, wherein the characteristic is a type of communication device the person is using to communicate with the machine. 11. A system as recited in claim 1, wherein the characteristic is a type of communication device the person is using to communication with the machine. 12. A method as recited in claim 1, wherein the characteristic is indicative of a noise level. 13. A method as recited in claim 12, wherein the characteristic is indicative of an acoustic noise level of the dialog. 14. A method as recited in claim 12, wherein the characteristic is indicative of a signal noise level of the dialog. 15. A method as recited in claim 1, wherein the characteristic is a noise level of an acoustic environment in which the person is located. 16. A method as recited in claim 15, wherein the characteristic is a noise type of an acoustic environment in which the person is located. 17. A method as recited in claim 15, wherein the characteristic is the level of reverberance of an acoustic environment in which the person is located. 18. A system comprising: a front end to generate a set of features in response to speech from a person during a dialog with the person, wherein the person uses a communication device during a call communicate with the system via a communication channel; a set of models; a speech recognition engine to recognize the speech from the person based on the features and the models; a characteristic detector to detect a characteristic of the person, the communication device, the communication channel, or an environment in which the person is located during the dialog, wherein the characteristic does not uniquely identify the person, the communication device, or any user account; and a call routing unit to select a destination to which the call from the person should be routed, based on the detected characteristic, and not based on the meaning of any speech or the failure to recognize any speech during the dialog. 19. A method as recited in claim 18, wherein the characteristic is an approximate age of the person. 20. A system as recited in claim 18, wherein the characteristic is the gender of the person. 21. A system as recited in claim 18, wherein the characteristic is a type of speech being spoken by the person. 22. A system as recited in claim 18, wherein the characteristic is an emotional state of the person. 23. A system as recited in claim 18, wherein the characteristic is indicative of the truthfulness of speech of the person. 24. A system as recited in claim 18, wherein the characteristic is an acoustic characteristic. 25. A system as recited in claim 18, wherein the characteristic is indicative of a speech level of the dialog. 26. A system as recited in claim 18, wherein the characteristic is descriptive of a reason the person is experiencing an error. 27. A system as recited in claim 18, wherein the characteristic is indicative of a noise level. 28. A system as recited in claim 27, wherein the characteristic is indicative of an accoustic noise level of the dialog. 29. A system as recited in claim 27, wherein the characteristic is indicative of a signal noise level of the dialog. 30. A system as recited in claim 18, wherein the characteristic is a noise level of an accoustic environment in which the person is located. 31. A system as recited in claim 30, wherein the characteristic is a noise type of an acoustic environment in which the person is located. 32. A system as recited in claim 30, wherein the characteristic is the level of reverberance of an acoustic environment in which the person is located. 33. A method comprising: establishing a speech-based dialog between a person and a machine, wherein the person uses a communication device to communicate with the machine via a communication channel during the dialog; automatically detecting a characteristic during the dialog in real time, wherein the characteristic is a characteristic of the person, the communication device, the communication channel, or an environment in which the peson is located during the dialog, and wherein the characteristic does not uniquely identify the person, the communication device, or any user account; and dynamically customizing a call flow of the dialog for the person during the dialog, based on the detected characteristic, and not based on the meaning of any speech or the failure to recognize any speech during the dialog. 34. A method as recited in claim 33, wherein the characteristic is a characteristic of the person's speech during the dialog. 35. A method as recited in claim 33, wherein the characteristic is the person's gender or approximate age. 36. A method as recited in claim 33, wherein the characteristic is the person's emotional state. 37. A method as recited in claim 33, wherein the characteristic is an acoustic characteristic. 38. A method as recited in claim 33, wherein the characteristic is a type of communication device the person is using to communicate with the machine.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.