최소 단어 이상 선택하여야 합니다.
최대 10 단어까지만 선택 가능합니다.
SAI
다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
NTIS 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
DataON 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
Edison 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
Kafe 바로가기국가/구분 | United States(US) Patent 등록 |
---|---|
국제특허분류(IPC7판) |
|
출원번호 | US-0632713 (2017-06-26) |
등록번호 | US-10089984 (2018-10-02) |
발명자 / 주소 |
|
출원인 / 주소 |
|
대리인 / 주소 |
|
인용정보 | 피인용 횟수 : 0 인용 특허 : 525 |
A system and method for an integrated, multi-modal, multi-device natural language voice services environment may be provided. In particular, the environment may include a plurality of voice-enabled devices each having intent determination capabilities for processing multi-modal natural language inpu
A system and method for an integrated, multi-modal, multi-device natural language voice services environment may be provided. In particular, the environment may include a plurality of voice-enabled devices each having intent determination capabilities for processing multi-modal natural language inputs in addition to knowledge of the intent determination capabilities of other devices in the environment. Further, the environment may be arranged in a centralized manner, a distributed peer-to-peer manner, or various combinations thereof. As such, the various devices may cooperate to determine intent of multi-modal natural language inputs, and commands, queries, or other requests may be routed to one or more of the devices best suited to take action in response thereto.
1. A method of providing an integrated multi-modal, natural language voice services environment comprising one or more of an input device that receives a multi-modal natural language input comprising at least a natural language utterance and a non-voice input related to the natural language utteranc
1. A method of providing an integrated multi-modal, natural language voice services environment comprising one or more of an input device that receives a multi-modal natural language input comprising at least a natural language utterance and a non-voice input related to the natural language utterance, a first device, or one or more secondary devices, the method being implemented in the first device having one or more physical processors programmed with computer program instructions that, when executed by the one or more physical processors, program the first device to perform the method, the method comprising: obtaining, by the first device from the input device, the multi-modal natural language input;transcribing, by the first device, the natural language utterance;determining, by the first device, a preliminary intent prediction of the multi-modal natural language input based on the transcribed utterance and the non-voice input; andinvoking, by the first device, at least one action at one or more of the input device, the first device, or the one or more secondary devices based on the preliminary intent prediction. 2. The method of claim 1, wherein invoking the at least one action at one or more of the input device, the first device, or the one or more secondary devices comprises transmitting a request related to the multi-modal natural language input based on the preliminary intent prediction. 3. The method of claim 1, wherein the one or more secondary devices include at least a second device, the method further comprising: transmitting, by the first device, the multi-modal natural language input to the second device;receiving, by the first device from the second device, a second intent prediction of the multi-modal natural language input; anddetermining, by the first device, an intent of the multi-modal natural language input based on the preliminary intent prediction and the second intent prediction, wherein the at least one action is invoked based on the determined intent. 4. The method of claim 3, the method further comprising: determining, by the first device, processing capabilities associated with the one or more secondary devices; and selecting, by the first device, based on the processing capabilities associated with the one or more secondary devices, the second device to make the second intent prediction of the multi-modal natural language input. 5. The method of claim 4, the method further comprising: maintaining, by the first device, a constellation model that describes natural language resources, dynamic states, and intent determination capabilities associated with the input device and the one or more secondary devices, wherein the processing capabilities associated with the one or more secondary devices are determined based on the constellation model. 6. The method of claim 5, wherein the intent determination capabilities for a given one of the input device, the first device, or the one or more secondary devices are based on at least one of processing power, storage resources, natural language processing capabilities, or local knowledge. 7. The method of claim 3, the method further comprising: determining, by the first device, a domain relating to the multi-modal natural language input; and selecting, by the first device, based on the domain, the second device to make the second intent prediction of the multi-modal natural language input. 8. The method of claim 7, wherein the one or more secondary devices are associated with different domains, the second device is associated with the domain, and the different domains comprise the domain. 9. The method of claim 1, the method further comprising: communicating, by the first device, the multi-modal natural language input to each of the one or more secondary devices, wherein each of the one or more secondary devices determine an intent of the multi-modal natural language input received at the input device using local intent determination capabilities;receiving, by the first device, the intent determined by each of the secondary devices; andarbitrating, by the first device, among the intent determinations of the secondary devices to determine the intent of the multi-modal natural input, wherein the at least one action is invoked based on the arbitrated intent determinations. 10. The method of claim 1, wherein the input device initially received the multi-modal natural language input. 11. A system for processing a multi-modal natural language input, the system comprising: an input device that receives a multi-modal natural language input comprising at least a natural language utterance and a non-voice input related to the natural language utterance; one or more secondary devices; anda first device having one or more physical processors programmed with computer program instructions that, when executed by the one or more physical processors, program the first device to:obtain, from the input device, the multi-modal natural language input; transcribe the natural language utterance;determine a preliminary intent prediction of the multi-modal natural language input based on the transcribed utterance and the non-voice input; andinvoke at least one action at one or more of the input device, the first device, or the one or more secondary devices based on the preliminary intent prediction. 12. The system of claim 11, wherein to invoke the at least one action at one or more of the input device, the first device, or the one or more secondary devices, the first device is further programmed to: transmit a request related to the multi-modal natural language input based on the preliminary intent prediction. 13. The system of claim 11, wherein the one or more secondary devices include at least a second device, and wherein the first device is further programmed to: transmit the multi-modal natural language input to the second device; receive, from the second device, a second intent prediction of the multi-modal natural language input; and determine an intent of the multi-modal natural language input based on the preliminary intent prediction and the second intent prediction, wherein the at least one action is invoked based on the determined intent. 14. The system of claim 13, wherein the first device is further programmed to: determine processing capabilities associated with the one or more secondary devices; and select based on the processing capabilities associated with the one or more secondary devices, the second device to make the second intent prediction of the multi-modal natural language input. 15. The system of claim 14, wherein the first device is further programmed to: maintain a constellation model that describes natural language resources, dynamic states, and intent determination capabilities associated with the input device and the one or more secondary devices, wherein the processing capabilities associated with the one or more secondary devices are determined based on the constellation model. 16. The system of claim 15, wherein the intent determination capabilities for a given one of the input device, the first device, or the one or more secondary devices are based on at least one of processing power, storage resources, natural language processing capabilities, or local knowledge. 17. The system of claim 13, wherein the first device is further programmed to: determine a domain relating to the multi-modal natural language input; and select, based on the domain, the second device to make the second intent prediction of the multi-modal natural language input. 18. The system of claim 17, wherein the one or more secondary devices are associated with different domains, the second device is associated with the domain, and the different domains comprise the domain. 19. The system of claim 11, wherein the first device is further programmed to: communicate the multi-modal natural language input to each of the one or more secondary devices, wherein each of the one or more secondary devices determine an intent of the multi-modal natural language input received at the input device using local intent determination capabilities;receive the intent determined by each of the secondary devices; and arbitrate among the intent determinations of the secondary devices to determine the intent of the multi-modal natural input, wherein the at least one action is invoked based on the arbitrated intent determinations. 20. The system of claim 11, wherein the input device initially received the multi-modal natural language input.
해당 특허가 속한 카테고리에서 활용도가 높은 상위 5개 콘텐츠를 보여줍니다.
더보기 버튼을 클릭하시면 더 많은 관련자료를 살펴볼 수 있습니다.
IPC | Description |
---|---|
A | 생활필수품 |
A62 | 인명구조; 소방(사다리 E06C) |
A62B | 인명구조용의 기구, 장치 또는 방법(특히 의료용에 사용되는 밸브 A61M 39/00; 특히 물에서 쓰이는 인명구조 장치 또는 방법 B63C 9/00; 잠수장비 B63C 11/00; 특히 항공기에 쓰는 것, 예. 낙하산, 투출좌석 B64D; 특히 광산에서 쓰이는 구조장치 E21F 11/00) |
A62B-1/08 | .. 윈치 또는 풀리에 제동기구가 있는 것 |
내보내기 구분 |
|
---|---|
구성항목 |
관리번호, 국가코드, 자료구분, 상태, 출원번호, 출원일자, 공개번호, 공개일자, 등록번호, 등록일자, 발명명칭(한글), 발명명칭(영문), 출원인(한글), 출원인(영문), 출원인코드, 대표IPC 관리번호, 국가코드, 자료구분, 상태, 출원번호, 출원일자, 공개번호, 공개일자, 공고번호, 공고일자, 등록번호, 등록일자, 발명명칭(한글), 발명명칭(영문), 출원인(한글), 출원인(영문), 출원인코드, 대표출원인, 출원인국적, 출원인주소, 발명자, 발명자E, 발명자코드, 발명자주소, 발명자 우편번호, 발명자국적, 대표IPC, IPC코드, 요약, 미국특허분류, 대리인주소, 대리인코드, 대리인(한글), 대리인(영문), 국제공개일자, 국제공개번호, 국제출원일자, 국제출원번호, 우선권, 우선권주장일, 우선권국가, 우선권출원번호, 원출원일자, 원출원번호, 지정국, Citing Patents, Cited Patents |
저장형식 |
|
메일정보 |
|
안내 |
총 건의 자료가 검색되었습니다. 다운받으실 자료의 인덱스를 입력하세요. (1-10,000) 검색결과의 순서대로 최대 10,000건 까지 다운로드가 가능합니다. 데이타가 많을 경우 속도가 느려질 수 있습니다.(최대 2~3분 소요) 다운로드 파일은 UTF-8 형태로 저장됩니다. ~ |
Copyright KISTI. All Rights Reserved.
AI-Helper는 오픈소스 모델을 사용합니다. 사용하고 있는 오픈소스 모델과 라이센스는 아래에서 확인할 수 있습니다.
AI-Helper uses Open Source Models. You can find the source code of these open source models, along with applicable license information below. (helpdesk@kisti.re.kr)
OpenAI의 API Key를 브라우저에 등록하여야 ChatGPT 모델을 사용할 수 있습니다.
등록키는 삭제 버튼을 누르거나, PDF 창을 닫으면 삭제됩니다.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.