최소 단어 이상 선택하여야 합니다.
최대 10 단어까지만 선택 가능합니다.
다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
NTIS 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
DataON 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
Edison 바로가기다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
Kafe 바로가기국가/구분 | United States(US) Patent 등록 |
---|---|
국제특허분류(IPC7판) |
|
출원번호 | US-0298720 (2014-06-06) |
등록번호 | US-9582608 (2017-02-28) |
발명자 / 주소 |
|
출원인 / 주소 |
|
대리인 / 주소 |
|
인용정보 | 피인용 횟수 : 6 인용 특허 : 2015 |
Methods, systems, and computer-readable media related to a technique for combining two or more aspects of predictive information for auto-completion of user input, in particular, user commands directed to an intelligent digital assistant. Specifically, predictive information based on (1) usage frequ
Methods, systems, and computer-readable media related to a technique for combining two or more aspects of predictive information for auto-completion of user input, in particular, user commands directed to an intelligent digital assistant. Specifically, predictive information based on (1) usage frequency, (2) usage recency, and (3) semantic information encapsulated in an ontology (e.g., a network of domains) implemented by the digital assistant, are integrated in a balanced and sensible way within a unified framework, such that a consistent ranking of all completion candidates across all domains may be achieved. Auto-completions are selected and presented based on the unified ranking of all completion candidates.
1. A method of providing cross-domain semantic ranking of complete input phrases for a digital assistant, comprising: receiving a training corpus comprising a collection of complete input phrases that span a plurality of semantically distinct domains;for each of a plurality of distinct words present
1. A method of providing cross-domain semantic ranking of complete input phrases for a digital assistant, comprising: receiving a training corpus comprising a collection of complete input phrases that span a plurality of semantically distinct domains;for each of a plurality of distinct words present in the collection of complete input phrases, calculating a respective word indexing power across the plurality of domains based on a respective normalized entropy for said word, wherein the respective normalized entropy is based on a total number of domains in which said word appears and how representative said word is for each of the plurality of domains;for each complete input phrase in the collection of complete input phrases, calculating a respective phrase indexing power across the plurality of domains based on an aggregation of the respective word indexing powers of all constituent words of said complete input phrase;obtaining respective domain-specific usage frequencies of the complete input phrases in the training corpus; andgenerating a cross-domain ranking of the collection of complete input phrases based at least on the respective phrase indexing powers of the complete input phrases and the respective domain-specific usage frequencies of the complete input phrases. 2. The method of claim 1, further comprising: providing the cross-domain ranking of the collection of complete input phrases to a user device, wherein the user device presents one or more auto-completion candidates in response to an initial user input in accordance with at least the cross-domain ranking of the collection of complete input phrases. 3. The method of claim 1, wherein calculating the respective word indexing power across the plurality of domains for each word wi of the plurality of distinct words further comprises: calculating the respective normalized entropy εi; for the word wi based on a respective formula ɛi=-1logK∑k=1Kci,ktilogci,kti, wherein K is a total number of domains in the plurality of domains, ci,k is a total number of times wi occurs in a domain dk of the plurality of domains, and ti=Σkci,k is a total number of times wi occurs in the collection of complete input phrases, and wherein the respective word indexing power of the word wi is (1−εi). 4. The method of claim 1, wherein calculating the respective phrase indexing power across the plurality of domains for each complete input phrase Pj of the collection of complete input phrases further comprises: distinguishing template words from normal words in the complete input phrase Pj, a template word being a word that is used to represent a respective category of normal words in a particular complete input phrase and that is substituted by one or more normal words when provided as an input to the digital assistant by a user;calculating the respective phrase indexing power, for the complete input phrase Pj based on a respective formula μj=bTnT(j)+1nN(j)+nT(j)[∑i=1nN(j)(1-ɛi)+∑i=1nT(j)(1-ɛi)], wherein nN(j) is a total number of normal words present in the complete input phrase Pj, nT(j) is a total number of template words present in the complete input phrase Pj, (1−εi) is the respective word indexing power of each word wi, and bT is a respective template bias multiplier used to calculate the weight bias bTnT(j) for the input phrase Pj. 5. The method of claim 1, wherein generating the cross-domain ranking of the collection of complete input phrases further comprises: calculating a respective integrated ranking score Rj for each complete input phrase Pj of the collection of complete input phrases based on a respective formula Rj=ωvvj+ωμμjωv+ωμ, wherein vj is a respective frequency of the complete input phrase Pj that has been normalized for cross-domain comparison, uj is a respective phrase indexing power of the complete input phrase Pj across the plurality of domains, ωv and ωμ are relative weights given to domain-specific frequency and cross-domain indexing power in the ranking of the collection of complete input phrases; and generating the cross-domain ranking of the collection of complete input phrases based on the respective integrated ranking scores for the collection of complete input phrases. 6. The method of claim 1, further comprising: receiving the initial user input from a user;identifying, from the collection of complete input phrases, a subset of complete input phrases that each begins with the initial user input;ranking the subset of complete input phrases in accordance with the cross-domain ranking of the collection of complete input phrases;selecting a predetermined number of unique relevant domains based on respective domains associated with each of the subset of complete input phrases; andselecting at least one top-ranked input phrase from each of the unique relevant domains as one of the auto-completion candidates to be presented to the user. 7. The method of claim 2, wherein presenting one or more auto-completion candidates further comprises: displaying a first portion of a first auto-completion candidate of the one or more auto-completion candidates, wherein the first portion of the first auto-completion candidate precedes or ends at a respective template word in the first auto-completion candidate;receiving a subsequent user input specifying one or more normal words corresponding to the respective template word; andin response to receiving the subsequent user input, displaying a second portion of the first auto-completion candidate succeeding the respective template word in the first auto-completion candidate. 8. A non-transitory computer-readable medium having instructions stored thereon, the instructions, when executed by one or more processors, cause the processors to perform operations comprising: receiving a training corpus comprising a collection of complete input phrases that span a plurality of semantically distinct domains;for each of a plurality of distinct words present in the collection of complete input phrases, calculating a respective word indexing power across the plurality of domains based on a respective normalized entropy for said word, wherein the respective normalized entropy is based on a total number of domains in which said word appears and how representative said word is for each of the plurality of domains;for each complete input phrase in the collection of complete input phrases, calculating a respective phrase indexing power across the plurality of domains based on an aggregation of the respective word indexing powers of all constituent words of said complete input phrase;obtaining respective domain-specific usage frequencies of the complete input phrases in the training corpus; andgenerating a cross-domain ranking of the collection of complete input phrases based at least on the respective phrase indexing powers of the complete input phrases and the respective domain-specific usage frequencies of the complete input phrases. 9. The computer-readable medium of claim 8, wherein the operations further comprise: providing the cross-domain ranking of the collection of complete input phrases to a user device, wherein the user device presents one or more auto-completion candidates in response to an initial user input in accordance with at least the cross-domain ranking of the collection of complete input phrases. 10. The computer-readable medium of claim 9, wherein calculating the respective word indexing power across the plurality of domains for each word w1 of the plurality of distinct words further comprises: calculating the respective normalized entropy εi for the word wi based on a respective ɛi=-1logK∑k=1Kci,ktilogci,kti, formula wherein K is a total number of domains in the plurality of domains, ci,k is a total number of times wi occurs in a domain dk of the plurality of domains, and ti=Σkci,k is a total number of times wi occurs in the collection of complete input phrases, and wherein the respective word indexing power of the word wi is (1−εi). 11. The computer-readable medium of claim 9, wherein calculating the respective phrase indexing power across the plurality of domains for each complete input phrase Pj of the collection of complete input phrases further comprises: distinguishing template words from normal words in the complete input phrase Pj, a template word being a word that is used to represent a respective category of normal words in a particular complete input phrase and that is substituted by one or more normal words when provided as an input to a digital assistant by a user;calculating the respective phrase indexing power μj for the complete input phrase Pj based on a respective formula μj=bTnT(j)+1nN(j)+nT(j)[∑i=1nN(j)(1-ɛi)+∑i=1nT(j)(1-ɛi)], wherein nN(j) is a total number of normal words present in the complete input phrase Pj, nT(j) is a total number of template words present in the complete input phrase Pj, (1−εi) is the respective word indexing power of each word wi, and bT is a respective template bias multiplier used to calculate the weight bias bTnT(j) for the input phrase P1. 12. The computer-readable medium of claim 9, wherein generating the cross-domain ranking of the collection of complete input phrases further comprises: calculating a respective integrated ranking score Rj for each complete input phrase Pj of the collection of complete input phrases based on a respective formula Rj=ωvvj+ωμμjωv+ωμ, wherein vj is a respective frequency of the complete input phrase Pj that has been normalized for cross-domain comparison, μj is a respective phrase indexing power of the complete input phrase Pj across the plurality of domains, ωv and ωμ are relative weights given to domain-specific frequency and cross-domain indexing power in the ranking of the collection of complete input phrases; and generating the cross-domain ranking of the collection of complete input phrases based on the respective integrated ranking scores for the collect ion of complete input phrases. 13. The computer-readable medium of claim 8, wherein the operations further comprise: receiving the initial user input from a user;identifying, from the collection of complete input phrases, a subset of complete input phrases that each begins with the initial user input;ranking the subset of complete input phrases in accordance with the cross-domain ranking of the collection of complete input phrases;selecting a predetermined number of unique relevant domains based on respective domains associated with each of the subset of complete input phrases; andselecting at least one top-ranked input phrase from each of the unique relevant domains as one of the auto-completion candidates to be presented to the user. 14. The computer-readable medium of claim 9, wherein presenting one or more auto-completion candidates further comprises: displaying a first portion of a first auto-completion candidate of the one or more auto-completion candidates, wherein the first portion of the first auto-completion candidate precedes or ends at a respective template word in the first auto-completion candidate;receiving a subsequent user input specifying one or more normal words corresponding to the respective template word; andin response to receiving the subsequent user input, displaying a second portion of the first auto-completion candidate succeeding the respective template word in the first auto-completion candidate. 15. The computer-readable medium of claim 14, wherein the operations further comprise: displaying one or more suggestions corresponding to the respective template word based on a user-specific vocabulary comprising at least one of a plurality of proper nouns associated with the user, wherein the subsequent user input is a selection of the one or more displayed suggestions. 16. A system, comprising: one or more processors; andmemory having instructions stored thereon, the instructions, when executed by the one or more processors, cause the processors to perform operations comprising:receiving a training corpus comprising a collection of complete input phrases that span a plurality of semantically distinct domains;for each of a plurality of distinct words present in the collection of complete input phrases, calculating a respective word indexing power across the plurality of domains based on a respective normalized entropy for said word, wherein the respective normalized entropy is based on a total number of domains in which said word appears and how representative said word is for each of the plurality of domains;for each complete input phrase in the collection of complete input phrases, calculating a respective phrase indexing power across the plurality of domains based on an aggregation of the respective word indexing powers of all constituent words of said complete input phrase;obtaining respective domain-specific usage frequencies of the complete input phrases in the training corpus; andgenerating a cross-domain ranking of the collection of complete input phrases based at least on the respective phrase indexing powers of the complete input phrases and the respective domain-specific usage frequencies of the complete input phrases. 17. The system of claim 16, wherein the operations further comprise: providing the cross-domain ranking of the collection of complete input phrases to a user device, wherein the user device presents one or more auto-completion candidates in response to an initial user input in accordance with at least the cross-domain ranking of the collection of complete input phrases. 18. The system of claim 16, wherein calculating the respective word indexing power across the plurality of domains for each word wi of the plurality of distinct words further comprises: calculating the respective normalized entropy εi for the word wi based on a respective formula ɛi=-1logK∑k=1Kci,ktilogci,kti, wherein K is a total number of domains in the plurality of domains, ci,k is a total number of times wi occurs in a domain dk of the plurality of domains, and ti=Σkci,k is a total number of times wi occurs in the collection of complete input phrases, and wherein the respective word indexing power of the word wi is (1−εi). 19. The system of claim 16, wherein calculating the respective phrase indexing power across the plurality of domains for each complete input phrase Pj of the collection of complete input phrases further comprises: distinguishing template words from normal words in the complete input phrase Pj, a template word being a word that is used to represent a respective category of normal words in a particular complete input phrase and that is substituted by one or more normal words when provided as an input to a digital assistant by a user;calculating the respective phrase indexing power μj for the complete input phrase Pj based on a respective formula μj=bTnT(j)+1nN(j)+nT(j)[∑i=1nN(j)(1-ɛi)+∑i=1nT(j)(1-ɛi)], wherein nN (j) is a total number of normal words present in the complete input phrase Pj, nT(j) is a total number of template words present in the complete input phrase Pj, (1−εi) is the respective word indexing power of each word wi, and bT is a respective template bias multiplier used to calculate the weight bias bTnT(j) for the input phrase Pj. 20. The system of claim 16, wherein generating the cross-domain ranking of the collection of complete input phrases further comprises: calculating a respective integrated ranking score Rj for each complete input phrase Pj of the collection of complete input phrases based on a respective formula Rj=ωvvj+ωμμjωv+ωμ, wherein vj is a respective frequency of the complete input phrase Pj that has been normalized for cross-domain comparison, μj is a respective phrase indexing power of the complete input phrase Pj across the plurality of domains, ωv and ωμ are relative weights given to domain-specific frequency and cross-domain indexing power in the ranking of the collection of complete input phrases; and generating the cross-domain ranking of the collection of complete input phrases based on the respective integrated ranking scores for the collection of complete input phrases. 21. The system of claim 20, wherein the operations further comprise: for each complete input phrase Pj, normalizing the respective domain-specific frequency of said complete input phrase Pj by a maximum phrase count of a single input phrase observed in the training corpus. 22. The system of claim 21, wherein the operations further comprise: updating the respective rank of each complete input phrase Pj based on a user-specific recency bias bR, wherein the user-specific recency bias is based on a number of times that a particular user has used the complete input phrase Pj. 23. The system of claim 16, wherein the operations further comprise: receiving the initial user input from a user;identifying, from the collection of complete input phrases, a subset of complete input phrases that each begins with the initial user input;ranking the subset of complete input phrases in accordance with the cross-domain ranking of the collection of complete input phrases;selecting a predetermined number of unique relevant domains based on respective domains associated with each of the subset of complete input phrases; andselecting at least one top-ranked input phrase from each of the unique relevant domains as one of the auto-completion candidates to be presented to the user. 24. The system of claim 17, wherein presenting one or more auto-completion candidates further comprises: displaying a first portion of a first auto-completion candidate of the one or more auto-completion candidates, wherein the first portion of the first auto-completion candidate precedes or ends at a respective template word in the first auto-completion candidate;receiving a subsequent user input specifying one or more normal words corresponding to the respective template word; andin response to receiving the subsequent user input, displaying a second portion of the first auto-completion candidate succeeding the respective template word in the first auto-completion candidate. 25. The system of claim 24, wherein the operations further comprise: displaying one or more suggestions corresponding to the respective template word based on a user-specific vocabulary comprising at least one of a plurality of proper nouns associated with the user, wherein the subsequent user input is a selection of the one or more displayed suggestions. 26. The method of claim 4, wherein the respective template bias multiplier is a positive real number, and wherein the method further comprises: adjusting the respective template bias multiplier based on a performance evaluation of the cross-domain ranking in providing auto-completion candidates for user input. 27. The method of claim 4, wherein receiving a training corpus further comprises: collecting a plurality of user input phrases from a usage log of a digital assistant;identifying at least one template input phrase based on a common phrase pattern present in two or more of the plurality of user input phrases; andnormalizing the plurality of user input phrases by substituting at least one word in each of said two or more user input phrases with a respective template word representing a generalization of the at least one word.
Copyright KISTI. All Rights Reserved.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.