최소 단어 이상 선택하여야 합니다.
최대 10 단어까지만 선택 가능합니다.
다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
NTIS 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
DataON 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
Edison 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
Kafe 바로가기국가/구분 | United States(US) Patent 등록 |
---|---|
국제특허분류(IPC7판) |
|
출원번호 | US-0169348 (2016-05-31) |
등록번호 | US-10223066 (2019-03-05) |
발명자 / 주소 |
|
출원인 / 주소 |
|
대리인 / 주소 |
|
인용정보 | 피인용 횟수 : 0 인용 특허 : 2093 |
Systems and processes for proactive assistance based on dialog communication between devices are provided. In one example process, while voice communication between an electronic device and a second electronic device is established, a stream of audio data associated with the second electronic device
Systems and processes for proactive assistance based on dialog communication between devices are provided. In one example process, while voice communication between an electronic device and a second electronic device is established, a stream of audio data associated with the second electronic device can be received. In response to detecting a user input, a text representation of speech contained in a portion of the stream of audio data can be generated. The process can determine whether the text representation contains information corresponding to one of a plurality of types of information. In response to determining that the text representation contains information corresponding to one of a plurality of types of information, one or more tasks based on the information can be performed.
1. A non-transitory computer-readable medium storing instructions for providing proactive assistance based on dialog communication between devices, the instructions, when executed by one or more processors, cause the one or more processors to: upon establishing voice communication between an electro
1. A non-transitory computer-readable medium storing instructions for providing proactive assistance based on dialog communication between devices, the instructions, when executed by one or more processors, cause the one or more processors to: upon establishing voice communication between an electronic device and a second electronic device, display a first user interface on a display of the electronic device;while the voice communication is established between the electronic device and the second electronic device: receive a stream of audio data from the second electronic device;automatically store a predetermined duration of the stream of audio data; detect a user input separate from the stream of audio data; andin response to detecting the user input: generate a text representation of speech contained in the predetermined duration of the stream of audio data;determine whether the text representation contains information corresponding to one of a plurality of types of information; andin response to determining that the text representation contains information corresponding to one of a plurality of types of information, perform one or more tasks based on the information, wherein the one or more tasks include displaying, on the display, a selectable second user interface overlapping at least part of the displayed first user interface, wherein the second user interface contains the information, wherein the first user interface and the second user interface are displayed while the electronic device is in a locked state. 2. The computer-readable medium of claim 1, wherein the user input is detected by a physical sensor of the electronic device. 3. The computer-readable medium of claim 1, wherein a proximity sensor of the electronic device is positioned less than a threshold distance from a surface, and wherein the user input includes changing a position of the electronic device such that the proximity sensor is positioned greater than the threshold distance from the surface. 4. The computer-readable medium of claim 1, wherein the user input is detected by an accelerometer of the electronic device, and wherein the user input includes positioning the electronic device for a user to view a display of the electronic device. 5. The computer-readable medium of claim 1, wherein the user input includes selecting an affordance on the electronic device. 6. The computer-readable medium of claim 1, wherein the user input includes a user utterance, and wherein detecting the user input includes detecting one or more trigger words in the user utterance. 7. The computer-readable medium of claim 1, wherein the instructions further cause the one or more processors to: determine whether the information is complete;in response to determining that the information is not complete, generate a text representation of speech contained in a portion of the stream of audio data that is different from the predetermined duration of the stream of audio data; andobtain second information from the portion of the stream of audio data, wherein the one or more tasks are based on the second information. 8. The computer-readable medium of claim 1, wherein the information includes a telephone number, and wherein the instructions further cause the one or more processors to: in response to detecting a user selection of the displayed second user interface, initiate a voice call based on the telephone number. 9. The computer-readable medium of claim 1, wherein the information includes a telephone number, and wherein the instructions further cause the one or more processors to: in response to detecting a user selection of the telephone number displayed on the second user interface, store the telephone number in association with an address book of the electronic device. 10. The computer-readable medium of claim 1, wherein the information includes an email address, and wherein the instructions further cause the one or more processors to: in response to detecting a user selection of the displayed second user interface, initiate a composition of an email message, wherein a recipient of the email message is based on the email address. 11. The computer-readable medium of claim 1, wherein the information includes a location, and wherein the one or more tasks include displaying a map indicating the location. 12. The computer-readable medium of claim 1, wherein the user input includes a gesture input by a user. 13. The computer-readable medium of claim 1, further comprising instructions for: automatically storing at least a first portion of the stream of audio data, the first portion received prior to detecting the user input; andautomatically storing at least a second portion of the stream of audio data, the second portion received after detecting the user input. 14. The computer-readable medium of claim 1, wherein the stored predetermined duration of the stream of audio data is overwritten after a second predetermined duration. 15. The computer-readable medium of claim 1, wherein instructions to automatically store a predetermined duration of the stream of audio data further comprises instructions for: automatically storing only the predetermined duration of the stream of audio data, wherein the predetermined duration includes only the most recent N seconds of the received stream of audio data. 16. The computer-readable medium of claim 1, wherein the instructions further cause the processors to: continue display of the second user interface for a predetermined duration after detecting that the voice communication between the electronic device and the second electronic device has ended. 17. The computer-readable medium of claim 1, wherein the instructions further cause the one or more processors to: cease display of the second user interface at a predetermined time after detecting that the voice communication between the electronic device and the second electronic device has ended. 18. The computer-readable medium of claim 17, wherein the information corresponding to one of a plurality of types of information is purged from a memory of the electronic device after the predetermined duration. 19. The computer-readable medium of claim 1, wherein the instructions further cause the one or more processors to: cease display of the first user interface in response to detecting that the voice communication between the electronic device and the second electronic device has ended. 20. The computer-readable medium of claim 1, wherein the instructions further cause the one or more processors to: detect a user selection of the selectable second user interface;in response to detecting a user selection of the selectable second user interface:display, on the display, a prompt to confirm performance of at least one additional action; andperform the at least one additional action after the voice communication has ended. 21. A method for providing proactive assistance based on dialog communication between devices, the method comprising: at an electronic device:upon establishing voice communication between the electronic device and a second electronic device, display a first user interface on a display of the electronic device;while the voice communication is established between the electronic device and the second electronic device: receiving a stream of audio data from the second electronic device;automatically store a predetermined duration of the stream of audio data; detecting a user input separate from the stream of audio data; andin response to detecting the user input: generating a text representation of speech contained in the predetermined duration of the stream of audio data;determining whether the text representation contains information corresponding to one of a plurality of types of information; andin response to determining that the text representation contains information corresponding to one of a plurality of types of information, performing one or more tasks based on the information, wherein the one or more tasks include displaying, on the display, a selectable second user interface overlapping at least part of the displayed first user interface, wherein the second user interface contains the information, wherein the first user interface and the second user interface are displayed while the electronic device is in a locked state. 22. The method of claim 21, wherein a proximity sensor of the electronic device is positioned less than a threshold distance from a surface, and wherein the user input includes changing a position of the electronic device such that the proximity sensor is positioned greater than the threshold distance from the surface. 23. The method of claim 21, wherein the user input includes selecting an affordance on the electronic device. 24. The method of claim 21, wherein the user input includes a user utterance, and wherein detecting the user input includes detecting one or more trigger words in the user utterance. 25. The method of claim 21, further comprising: determining whether the information is complete;in response to determining that the information is not complete, generating a text representation of speech contained in a portion of the stream of audio data that is different from the predetermined duration of the stream of audio data; andobtaining second information from the portion of the stream of audio data, wherein the one or more tasks are based on the second information. 26. The method of claim 21, wherein the information includes a telephone number, and further comprising: in response to detecting a user selection of the displayed second user interface, initiating a voice call based on the telephone number. 27. The method of claim 21, further comprising: continuing display of the second user interface for a predetermined duration after detecting that the voice communication between the electronic device and the second electronic device has ended. 28. The method of claim 21, further comprising: ceasing display of the second user interface at a predetermined time after detecting that the voice communication between the electronic device and the second electronic device has ended. 29. The method of claim 28, wherein the information corresponding to one of a plurality of types of information is purged from a memory of the electronic device after the predetermined duration. 30. The method of claim 21, further comprising: ceasing display of the first user interface in response to detecting that the voice communication between the electronic device and the second electronic device has ended. 31. The method of claim 21, further comprising: detecting a user selection of the selectable second user interface;in response to detecting a user selection of the selectable second user interface: displaying, on the display, a prompt to confirm performance of at least one additional action; andperforming the at least one additional action after the voice communication has ended. 32. The method of claim 21, wherein the stored predetermined duration of the stream of audio data is overwritten after a second predetermined duration. 33. An electronic device for providing proactive assistance based on dialog communication between devices, the electronic device comprising: one or more processors; andmemory having instructions stored thereon, the instructions, when executed by the one or more processors, cause the one or more processors to: upon establishing voice communication between the electronic device and a second electronic device, display a first user interface on a display of the electronic device;while the voice communication is established between the electronic device and the second electronic device: receive a stream of audio data from the second electronic device;automatically store a predetermined duration of the stream of audio data; detect a user input separate from the stream of audio data; andin response to detecting the user input: generate a text representation of speech contained in the predetermined duration of the stream of audio data;determine whether the text representation contains information corresponding to one of a plurality of types of information; andin response to determining that the text representation contains information corresponding to one of a plurality of types of information, perform one or more tasks based on the information, wherein the one or more tasks include displaying, on the display, a selectable second user interface overlapping at least part of the displayed first user interface, wherein the second user interface contains the information, wherein the first user interface and the second user interface are displayed while the electronic device is in a locked state. 34. The device of claim 33, wherein a proximity sensor of the electronic device is positioned less than a threshold distance from a surface, and wherein the user input includes changing a position of the electronic device such that the proximity sensor is positioned greater than the threshold distance from the surface. 35. The device of claim 33, wherein the user input includes selecting an affordance on the electronic device. 36. The device of claim 33, wherein the user input includes a user utterance, and wherein detecting the user input includes detecting one or more trigger words in the user utterance. 37. The device of claim 33, wherein the instructions further cause the one or more processors to: determine whether the information is complete;in response to determining that the information is not complete, generate a text representation of speech contained in a portion of the stream of audio data that is different from the predetermined duration of the stream of audio data; andobtain second information from the portion of the stream of audio data, wherein the one or more tasks are based on the second information. 38. The device of claim 33, wherein the information includes a telephone number, wherein the instructions further cause the one or more processors to: in response to detecting a user selection of the displayed second user interface, initiate a voice call based on the telephone number. 39. The device of claim 33, wherein the instructions further cause the one or more processors to: continue display of the second user interface for a predetermined duration after detecting that the voice communication between the electronic device and the second electronic device has ended. 40. The device of claim 33, wherein the instructions further cause the one or more processors to: ceasing display of the second user interface at a predetermined time after detecting that the voice communication between the electronic device and the second electronic device has ended. 41. The device of claim 40, wherein the information corresponding to one of a plurality of types of information is purged from a memory of the electronic device after the predetermined duration. 42. The device of claim 33, wherein the instructions further cause the one or more processors to: cease display of the first user interface in response to detecting that the voice communication between the electronic device and the second electronic device has ended. 43. The device of claim 33, wherein the instructions further cause the one or more processors to: detect a user selection of the selectable second user interface;in response to detecting a user selection of the selectable second user interface: display, on the display, a prompt to confirm performance of at least one additional action; andperform the at least one additional action after the voice communication has ended. 44. The device of claim 33, wherein the stored predetermined duration of the stream of audio data is overwritten after a second predetermined duration.
Copyright KISTI. All Rights Reserved.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.