US20030115060A1 - System and interactive form filling with fusion of data from multiple unreliable information sources - Google Patents
System and interactive form filling with fusion of data from multiple unreliable information sources Download PDFInfo
- Publication number
- US20030115060A1 US20030115060A1 US10/245,144 US24514402A US2003115060A1 US 20030115060 A1 US20030115060 A1 US 20030115060A1 US 24514402 A US24514402 A US 24514402A US 2003115060 A1 US2003115060 A1 US 2003115060A1
- Authority
- US
- United States
- Prior art keywords
- information
- user
- input
- prompt
- operable
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/24—Speech recognition using non-acoustical features
Definitions
- the present invention generally relates to speech recognition and particularly relates to automated form filling over a telephone system.
- the information fusion process should not trust the information sources to be reliable, and should intelligently use information from the multiple sources to constrain and supplement one another based on differences in reliability between sources and of particular information inputs.
- the fusion process should be able to incorporate a dialogue with a user to increase knowledge relating to reliability of information content, and/or gather additional information inputs.
- an automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources.
- An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources.
- a form filler is operable to fill an electronic form with the selected information.
- the form filling system of the present invention is advantageous over previous form filling systems in that it has knowledge relating to reliability of multiple information sources and is able to compare and select information content accordingly. It is further advantageous in its further aspects, wherein a prompt formulator is capable of formulating a prompt based on a comparison of the information content of the sources, and based on a level of knowledge relating to reliability of the source content, wherein the formulated prompt is designed to elicit a response from a user operable to increase knowledge of reliability relating to the multiple information sources.
- a resulting dialog with a user succeeds in filling out an electronic form while simultaneously decreasing effort on the user's part and increasing reliability of the form contents.
- the present invention is particularly suitable for use with electronic form filling over a telephone, wherein user input and accessible databases are unreliable information sources. Further areas of applicability of the present invention will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.
- FIG. 1 is an illustrated block diagram depicting a general implementation of the present invention with a telephone call center
- FIG. 2 is a flow chart depicting the method of the present invention
- FIG. 3 is a block diagram depicting a form filling system according to the present invention.
- FIG. 4 is a partial block and information flow diagram depicting a detailed implementation of the present invention with a telephone call center.
- the present invention is employed to perform form filling by fusing multimodal user input over a telephone with one or more user information databases.
- the form filling system 100 of FIG. 1 exemplifies one implementation of the preferred embodiment of the present invention to fill out a form requiring a user's name, address, and telephone number.
- a user one Mr. Baker, is initially prompted to speak his name into the telephone receiver as at 102 , and to spell his name via the telephone keypad as at 104 .
- the speech input 106 is communicated to an automatic speech recognizer 108
- the keypad entry 110 is communicated to an information database 112 of names indexed by predefined classes defined by the telephone keypad.
- a constraint list 114 of candidate names is generated from the keypad entry 110 and the information database 112 of names, and the generated constraint list 114 is further communicated to the speech recognizer 108 .
- the speech recognizer functions to recognize the speech input 106 by generating a plurality of speech recognition hypotheses, and then selects the N best generated hypotheses by comparing them to the constraint list 114 . Thus, if “Bater” were one of the original speech recognition hypotheses, it would be discarded based on it's absence from the constraint list.
- the information database 112 further serves as an information source having the names and addresses of most residents of the nation in which the user, Mr. Baker, is located, so frequency information 116 relating to the frequency with which names appear in the database can be communicated to rescoring module 118 used to further rescore the N best speech recognition hypotheses.
- rescoring module 118 used to further rescore the N best speech recognition hypotheses.
- the user is additionally prompted to enter his or her postal code, such as a zip code, and the received postal code 120 , the rescored speech recognition hypotheses, and the information database 112 are communicated to an information fuser 122 .
- the information fuser 122 selects information content of the information database 112 based on the rescored recognition hypotheses and the received zip code 120 .
- the information fuser 122 does not merely trust the database 112 , the rescored hypotheses, and the received zip code 120 to be accurate. Instead, the information fuser 122 recognizes and adjusts for several potential causes of unreliability.
- the data communicated to information fuser 122 may be unreliable for several reasons.
- the user's name may be spoken in such a way that the speech recognition system misrecognizes it.
- the customer may mistype his or her postal code.
- the customer database may be incomplete (address missing) or incorrect (name or address misspelled, information out of date).
- the information fuser employs an information selection strategy that selects the most reliable information available.
- the information fuser 122 selects the most reliable information available based on knowledge relating to reliability of various types of information sources, and based on a comparison of the information content from the different sources. For example, if the top-ranking speech recognition hypothesis does not match any name in the information database 112 having the received zip code 120 , but the second-highest ranking speech recognition hypothesis has only a slightly lower score than the top-ranking hypothesis and does match a name in the information database 112 having the received zip code 120 , then the information fuser can select the name indicated by the second-highest ranking speech recognition hypothesis and prompt the user for confirmation.
- the information fuser 122 can select the name and address matching the highest ranked speech recognition hypothesis and/or prompt the user to reenter the postal code or confirm whether the postal code is correct. Further, the information fuser may take a different approach by distinguishing between an entirely incorrect postal code and one that is only partially incorrect, and further consider the first two digits of the postal code (in the case of a zip code) more reliable than the last three digits of the postal code.
- Form filling system 100 preferably has an interviewer 123 for implementing a strategy to prompt the user for input increasing knowledge relating to reliability of information content.
- the interviewer 123 can be and/or incorporate a human agent to assist in confirming, correcting, selecting, and/or supplementing information.
- the interviewer 123 can be partially or wholly automated using a prompt formulator to request different inputs from the user in different situations.
- information fuser 122 can select, deselect, and/or reselect information content based on the increased knowledge relating to the reliability of the information content, and the user responses constitute new information sources to be used in the fusion process.
- the prompt formulator can preferably ask the user to supply supplemental information for fields of the form 124 for which reliable information cannot be obtained with the information sources currently available.
- Form filling system 100 also has a form filler for filling appropriate fields of the electronic form 124 with the selected information content.
- the method 200 of the present invention is illustrated in FIG. 2, and begins at 202 . Thence, the method 200 proceeds to step 203 , wherein information inputs are sought, preferably by initially prompting a user for specific inputs where a user is an applicable information source. Thence, the method proceeds to step 204 , wherein multiple information inputs are received from multiple information sources. In addition to information inputs provided by a user, these information inputs may also include data from an information database, and/or additional data such as that provided by caller ID or a biometric (measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.) of a user.
- biometric measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.
- step 206 information content is selected from one or more of the information sources based on a comparison of source contents and knowledge relating to reliability of the information sources.
- the knowledge of reliability of a source relates to the type of source, and thus is prior knowledge.
- the knowledge relating to reliability of an information source and/or specific information contents stems from a comparison of the contents of the information sources.
- content of an information source of a reliable type may be deemed less reliable based on comparison with content of another information source when the information content conflicts.
- content of an information source of an unreliable type may be deemed more reliable based on comparison with content of another information source when the information content matches. This process can be used to identify more and less reliable portions of information content within one or more information sources.
- step 208 an electronic form is filled with the selected information. If the form is deemed reliably completed as at 210 , then the method ends at 214 and a filled form has been generated. On the other hand, some or all of the selected information may be deemed insufficiently reliable, and/or a sufficient amount of fields of the form may not be deemed completed. In either of these latter cases, method 200 returns to step 203 .
- step 203 the user is prompted for additional information inputs, and the prompt is designed to elicit a response to increase knowledge of reliability relating to selectable information and/or gather supplemental information.
- the request for additional sources may be a request for confirmation of selected information, a request to reenter one of the user inputs, a request for a different information type, or a similar type of request.
- speech generation is preferred to communicate the request, especially over a telephone system.
- the prompt for additional information inputs the method 200 returns to step 204 , wherein the response is received. The method then proceeds again to step 206 , wherein new information content is selected and/or it is determined whether selected content is reliable based on the new information sources.
- the method 200 further proceeds to step 208 and fills the form based on the revised selection. If the newly filled form is deemed reliably completed at step 210 , then the method 200 ends at 216 . Otherwise, processing continues in a recursive fashion until a reliably completed form is obtained or the process is otherwise interrupted.
- a form filling system 300 of the present invention is more generally illustrated in FIG. 3, wherein a first information input 302 from a first information source and a second information input 304 from a second information source are received by input 306 and communicated to information fuser 308 .
- Information fuser 308 compares and selects information based on reliability of the information as determined based on the comparison and in accordance with predefined rules developed with regard to reliability of different types of information sources and/or information inputs.
- the selected information 310 is communicated to form filler 312 , which fills an electronic form 314 with the selected information to generate a filled form 316 and communicate it to an outside system via output 316 .
- Information fuser 308 mutually communicates with prompt formulator 320 , and prompt formulator 320 formulates a prompt 322 for additional information sources, if needed, based on reliability of the information sources, comparative and/or objective reliability of the available information inputs, and/or requirements for supplemental information content.
- Prompt formulator 320 further communicates its current state to information fuser 308 so that information fuser 308 is aware of the type of information input(s) requested and how to interpret its information content in view of the other information inputs.
- the formulated prompt 322 is communicated to a dialogue manager 324 that generates a prompt in a manner communicable to and understandable by a user, preferably by speech generation.
- the generated prompt is communicated to the user via output 328 .
- a response from the user constitutes an additional information input communicable to information fuser 308 via input 306 .
- multi-modal information inputs 400 from an information source corresponding to a user include a user speech input 402 and a user keypad entry 404 .
- the user speech input includes a spoken user name
- the user keypad entry includes a spelling and/or initials of the user name and a zip code for the user location.
- textual information inputs 406 from database information sources include nickname data input 408 from a nickname database and personal data input 410 from a personal information database. Use of nickname data input 408 in the present detailed implementation represents an improvement over the more simplified implementation previously discussed with reference to FIG. 1.
- additional data inputs 412 that also represent an improvement include telephone subscriber data 414 provided by a caller ID service.
- Other examples of additional data inputs 412 that can also be used include biometrics identifying a user and gathered, for example, by a handheld device the user employs to communicate with the call center. Further examples include an IP address for the user location, assuming the user communicates with the call center using a computer network.
- Form filling system 100 receives the multi-modal information inputs 400 , textual information inputs 406 , and additional data inputs 412 and uses the various information inputs to constrain and supplement one another according to their varying modalities, utilities, and levels of reliability.
- user speech input 402 is processed by speech recognizer 108 A to produce a plurality of speech recognition hypotheses 414 .
- user keypad entry 404 , additional data inputs 412 , nickname data input 408 , and personal data input 410 are all communicated to constraint list generator 416 .
- Constraint list generator 416 is adapted in the present implementation to generate a constraint list of candidate names by using a user keypad entry 404 containing a first and last name is used to access the nickname database and generate a plurality of first names based on the input first name, such that a keypad entry generating a constraint list containing “Bob” will also contain “Robert”, “Robby”, “Bobby”, and so on.
- the plurality of first names thus generated, a zip code from the user keypad entry, and any matching caller ID information are then used to access the personal information database, such as Phonedisc, containing names and addresses of all telephone subscribes in the United States, to generate a constraint list 114 that is communicated to N best hypotheses generator 108 B.
- the speech recognition hypotheses 414 are processes by N best hypotheses generator 108 B to generate an N best list of speech recognition hypotheses 418 .
- N best list of speech recognition hypotheses 418 is communicated to rescoring mechanism 118 , as is data from the personal information database pertaining to frequency of appearance of names in the personal information database.
- Rescoring mechanism 118 rescores the N best list of speech recognition hypotheses 418 to generate a list of rescored hypotheses 420 .
- Caller ID information and a postal code portion of the user keypad entry 404 may alternatively and/or additionally be used during this process to rescore hypotheses based on the frequency of names in the database having the corresponding postal code and/or names matching the caller ID information.
- the rescored hypotheses 420 , caller ID information, postal code, and personal data input 410 are communicated to the information fuser 122 and the prompt formulator 320 .
- the information fuser selects most reliable information from the personal data input 410 , caller ID information, and rescored hypotheses 420 , and the selected data 310 is communicated to the prompt formulator 320 and to form filler 312 .
- Form filler fills electronic form 314 to produce a filled form 316 that is partially or wholly filled with the selected data.
- the prompt formulator 320 determines whether more information is needed based on the selected data, the information available for selection, and requirements of the electronic form. If more information inputs are required, the prompt formulator formulates an appropriate prompt 322 and communicates a current state 422 to information fuser 122 . Otherwise, the form is deemed completed.
Abstract
Description
- This application is a continuation-in-part of U.S. patent application Ser. No. 10/022,023 filed on Dec. 13, 2001. The disclosure of the above application is incorporated herein by reference.
- The present invention generally relates to speech recognition and particularly relates to automated form filling over a telephone system.
- Automatic electronic form filling by a user, particularly over a telephone system, is a notoriously laborious and error prone process. Use of numerical keypad entries to attempt retrieval of user information is one existing process that proves to be only as reliable as completeness and correctness of the user information database and user knowledge of the required information and successful operation of the telephone keypad. Other applications, such as confirming availability and/or delivery of a product over the phone or on a networked computer system, suffer from similar problems.
- The key to improving the automated form filling process, by increasing reliability of information and decreasing effort on the part of a user, involves recognizing that information from multiple information sources can be fused in an intelligent manner. To be successful, the information fusion process should not trust the information sources to be reliable, and should intelligently use information from the multiple sources to constrain and supplement one another based on differences in reliability between sources and of particular information inputs. To be further successful, the fusion process should be able to incorporate a dialogue with a user to increase knowledge relating to reliability of information content, and/or gather additional information inputs.
- Automated form filling processes do not currently succeed in intelligently fusing information from multiple information sources based on knowledge relating to differences in reliability information form different sources, thereby simultaneously increasing reliability of form contents and decreasing effort on the part of the user. The need remains, therefore, for a solution to the problems associated with automated form filling as detailed above. The present invention provides such a solution.
- In accordance with the present invention, an automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources. An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources. A form filler is operable to fill an electronic form with the selected information.
- The form filling system of the present invention is advantageous over previous form filling systems in that it has knowledge relating to reliability of multiple information sources and is able to compare and select information content accordingly. It is further advantageous in its further aspects, wherein a prompt formulator is capable of formulating a prompt based on a comparison of the information content of the sources, and based on a level of knowledge relating to reliability of the source content, wherein the formulated prompt is designed to elicit a response from a user operable to increase knowledge of reliability relating to the multiple information sources. A resulting dialog with a user succeeds in filling out an electronic form while simultaneously decreasing effort on the user's part and increasing reliability of the form contents.
- The present invention is particularly suitable for use with electronic form filling over a telephone, wherein user input and accessible databases are unreliable information sources. Further areas of applicability of the present invention will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.
- The present invention will become more fully understood from the detailed description and the accompanying drawings, wherein:
- FIG. 1 is an illustrated block diagram depicting a general implementation of the present invention with a telephone call center;
- FIG. 2 is a flow chart depicting the method of the present invention;
- FIG. 3 is a block diagram depicting a form filling system according to the present invention; and
- FIG. 4 is a partial block and information flow diagram depicting a detailed implementation of the present invention with a telephone call center.
- The following description of the preferred embodiment(s) is merely exemplary in nature and is in no way intended to limit the invention, its application, or uses. In a preferred embodiment, the present invention is employed to perform form filling by fusing multimodal user input over a telephone with one or more user information databases. The
form filling system 100 of FIG. 1 exemplifies one implementation of the preferred embodiment of the present invention to fill out a form requiring a user's name, address, and telephone number. - According to form
filling system 100, a user, one Mr. Baker, is initially prompted to speak his name into the telephone receiver as at 102, and to spell his name via the telephone keypad as at 104. Thespeech input 106 is communicated to anautomatic speech recognizer 108, whereas thekeypad entry 110 is communicated to aninformation database 112 of names indexed by predefined classes defined by the telephone keypad. In turn, aconstraint list 114 of candidate names is generated from thekeypad entry 110 and theinformation database 112 of names, and the generatedconstraint list 114 is further communicated to thespeech recognizer 108. The speech recognizer functions to recognize thespeech input 106 by generating a plurality of speech recognition hypotheses, and then selects the N best generated hypotheses by comparing them to theconstraint list 114. Thus, if “Bater” were one of the original speech recognition hypotheses, it would be discarded based on it's absence from the constraint list. - The
information database 112 further serves as an information source having the names and addresses of most residents of the nation in which the user, Mr. Baker, is located, sofrequency information 116 relating to the frequency with which names appear in the database can be communicated to rescoringmodule 118 used to further rescore the N best speech recognition hypotheses. Thus, if “Baker” and “Bakes” are both present in the N best speech recognition hypotheses, they can be rescored to increase the ranking of “Baker” with respect to the ranking of “Bakes” based on a higher frequency of “Baker” in theinformation database 112 compared to the frequency of “Bakes”. - The user is additionally prompted to enter his or her postal code, such as a zip code, and the received
postal code 120, the rescored speech recognition hypotheses, and theinformation database 112 are communicated to aninformation fuser 122. The information fuser 122, in turn, selects information content of theinformation database 112 based on the rescored recognition hypotheses and the receivedzip code 120. The information fuser 122, however, does not merely trust thedatabase 112, the rescored hypotheses, and the receivedzip code 120 to be accurate. Instead, the information fuser 122 recognizes and adjusts for several potential causes of unreliability. - The data communicated to
information fuser 122 may be unreliable for several reasons. For example, the user's name may be spoken in such a way that the speech recognition system misrecognizes it. Also, the customer may mistype his or her postal code. Further, the customer database may be incomplete (address missing) or incorrect (name or address misspelled, information out of date). For these and similar reasons, the information fuser employs an information selection strategy that selects the most reliable information available. - The information fuser122 selects the most reliable information available based on knowledge relating to reliability of various types of information sources, and based on a comparison of the information content from the different sources. For example, if the top-ranking speech recognition hypothesis does not match any name in the
information database 112 having the receivedzip code 120, but the second-highest ranking speech recognition hypothesis has only a slightly lower score than the top-ranking hypothesis and does match a name in theinformation database 112 having the receivedzip code 120, then the information fuser can select the name indicated by the second-highest ranking speech recognition hypothesis and prompt the user for confirmation. Similarly, if the top-ranking speech recognition hypothesis does not match any name in theinformation database 112 having the receivedpostal code 120, but has a much higher score than the second-highest ranking speech recognition hypothesis, then theinformation fuser 122 can select the name and address matching the highest ranked speech recognition hypothesis and/or prompt the user to reenter the postal code or confirm whether the postal code is correct. Further, the information fuser may take a different approach by distinguishing between an entirely incorrect postal code and one that is only partially incorrect, and further consider the first two digits of the postal code (in the case of a zip code) more reliable than the last three digits of the postal code. -
Form filling system 100 preferably has an interviewer 123 for implementing a strategy to prompt the user for input increasing knowledge relating to reliability of information content. The interviewer 123 can be and/or incorporate a human agent to assist in confirming, correcting, selecting, and/or supplementing information. Alternatively, the interviewer 123 can be partially or wholly automated using a prompt formulator to request different inputs from the user in different situations. Accordingly,information fuser 122 can select, deselect, and/or reselect information content based on the increased knowledge relating to the reliability of the information content, and the user responses constitute new information sources to be used in the fusion process. Thus, the prompt formulator can preferably ask the user to supply supplemental information for fields of theform 124 for which reliable information cannot be obtained with the information sources currently available.Form filling system 100 also has a form filler for filling appropriate fields of theelectronic form 124 with the selected information content. - The
method 200 of the present invention is illustrated in FIG. 2, and begins at 202. Thence, themethod 200 proceeds tostep 203, wherein information inputs are sought, preferably by initially prompting a user for specific inputs where a user is an applicable information source. Thence, the method proceeds tostep 204, wherein multiple information inputs are received from multiple information sources. In addition to information inputs provided by a user, these information inputs may also include data from an information database, and/or additional data such as that provided by caller ID or a biometric (measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.) of a user. Thence, themethod 200 proceeds to step 206, wherein information content is selected from one or more of the information sources based on a comparison of source contents and knowledge relating to reliability of the information sources. In one aspect, the knowledge of reliability of a source relates to the type of source, and thus is prior knowledge. In another aspect, however, the knowledge relating to reliability of an information source and/or specific information contents stems from a comparison of the contents of the information sources. Thus, content of an information source of a reliable type may be deemed less reliable based on comparison with content of another information source when the information content conflicts. Similarly, content of an information source of an unreliable type may be deemed more reliable based on comparison with content of another information source when the information content matches. This process can be used to identify more and less reliable portions of information content within one or more information sources. - Once selection of information has taken place,
method 200 proceeds to step 208, wherein an electronic form is filled with the selected information. If the form is deemed reliably completed as at 210, then the method ends at 214 and a filled form has been generated. On the other hand, some or all of the selected information may be deemed insufficiently reliable, and/or a sufficient amount of fields of the form may not be deemed completed. In either of these latter cases,method 200 returns to step 203. - In
step 203, the user is prompted for additional information inputs, and the prompt is designed to elicit a response to increase knowledge of reliability relating to selectable information and/or gather supplemental information. Thus, the request for additional sources may be a request for confirmation of selected information, a request to reenter one of the user inputs, a request for a different information type, or a similar type of request. With this step, speech generation is preferred to communicate the request, especially over a telephone system. Following, the prompt for additional information inputs, themethod 200 returns to step 204, wherein the response is received. The method then proceeds again to step 206, wherein new information content is selected and/or it is determined whether selected content is reliable based on the new information sources. Themethod 200 further proceeds to step 208 and fills the form based on the revised selection. If the newly filled form is deemed reliably completed atstep 210, then themethod 200 ends at 216. Otherwise, processing continues in a recursive fashion until a reliably completed form is obtained or the process is otherwise interrupted. - A
form filling system 300 of the present invention is more generally illustrated in FIG. 3, wherein afirst information input 302 from a first information source and asecond information input 304 from a second information source are received byinput 306 and communicated toinformation fuser 308.Information fuser 308 compares and selects information based on reliability of the information as determined based on the comparison and in accordance with predefined rules developed with regard to reliability of different types of information sources and/or information inputs. The selectedinformation 310 is communicated to formfiller 312, which fills anelectronic form 314 with the selected information to generate a filledform 316 and communicate it to an outside system viaoutput 316. -
Information fuser 308 mutually communicates withprompt formulator 320, andprompt formulator 320 formulates a prompt 322 for additional information sources, if needed, based on reliability of the information sources, comparative and/or objective reliability of the available information inputs, and/or requirements for supplemental information content.Prompt formulator 320 further communicates its current state toinformation fuser 308 so thatinformation fuser 308 is aware of the type of information input(s) requested and how to interpret its information content in view of the other information inputs. The formulated prompt 322 is communicated to adialogue manager 324 that generates a prompt in a manner communicable to and understandable by a user, preferably by speech generation. The generated prompt is communicated to the user viaoutput 328. A response from the user constitutes an additional information input communicable toinformation fuser 308 viainput 306. - A detailed implementation of the present invention with a telephone call center is described with reference to FIG. 4. Therein,
multi-modal information inputs 400 from an information source corresponding to a user include auser speech input 402 and auser keypad entry 404. The user speech input includes a spoken user name, and the user keypad entry includes a spelling and/or initials of the user name and a zip code for the user location. Also,textual information inputs 406 from database information sources includenickname data input 408 from a nickname database andpersonal data input 410 from a personal information database. Use ofnickname data input 408 in the present detailed implementation represents an improvement over the more simplified implementation previously discussed with reference to FIG. 1. Further,additional data inputs 412 that also represent an improvement includetelephone subscriber data 414 provided by a caller ID service. Other examples ofadditional data inputs 412 that can also be used include biometrics identifying a user and gathered, for example, by a handheld device the user employs to communicate with the call center. Further examples include an IP address for the user location, assuming the user communicates with the call center using a computer network. These types of information inputs all assist to varying degrees in identifying the caller, and the form filling system of the present invention is adapted to use some or all of these types of information inputs accordingly. -
Form filling system 100 receives themulti-modal information inputs 400,textual information inputs 406, andadditional data inputs 412 and uses the various information inputs to constrain and supplement one another according to their varying modalities, utilities, and levels of reliability. For example,user speech input 402 is processed byspeech recognizer 108A to produce a plurality of speech recognition hypotheses 414. Also,user keypad entry 404,additional data inputs 412,nickname data input 408, andpersonal data input 410 are all communicated toconstraint list generator 416. -
Constraint list generator 416 is adapted in the present implementation to generate a constraint list of candidate names by using auser keypad entry 404 containing a first and last name is used to access the nickname database and generate a plurality of first names based on the input first name, such that a keypad entry generating a constraint list containing “Bob” will also contain “Robert”, “Robby”, “Bobby”, and so on. The plurality of first names thus generated, a zip code from the user keypad entry, and any matching caller ID information are then used to access the personal information database, such as Phonedisc, containing names and addresses of all telephone subscribes in the United States, to generate aconstraint list 114 that is communicated to Nbest hypotheses generator 108B. - The
speech recognition hypotheses 414 are processes by Nbest hypotheses generator 108B to generate an N best list of speech recognition hypotheses 418. One skilled in the art of speech recognition will recognize that an alternative embodiment may be realized with a word lattice. The N best list ofspeech recognition hypotheses 418 is communicated to rescoringmechanism 118, as is data from the personal information database pertaining to frequency of appearance of names in the personal information database.Rescoring mechanism 118 rescores the N best list ofspeech recognition hypotheses 418 to generate a list of rescoredhypotheses 420. Caller ID information and a postal code portion of theuser keypad entry 404 may alternatively and/or additionally be used during this process to rescore hypotheses based on the frequency of names in the database having the corresponding postal code and/or names matching the caller ID information. - The rescored
hypotheses 420, caller ID information, postal code, andpersonal data input 410 are communicated to theinformation fuser 122 and theprompt formulator 320. The information fuser selects most reliable information from thepersonal data input 410, caller ID information, and rescoredhypotheses 420, and the selecteddata 310 is communicated to theprompt formulator 320 and to formfiller 312. Form filler fillselectronic form 314 to produce a filledform 316 that is partially or wholly filled with the selected data. Meanwhile, theprompt formulator 320 determines whether more information is needed based on the selected data, the information available for selection, and requirements of the electronic form. If more information inputs are required, the prompt formulator formulates anappropriate prompt 322 and communicates acurrent state 422 toinformation fuser 122. Otherwise, the form is deemed completed. - The description of the invention is merely exemplary in nature and, thus, variations that do not depart from the gist of the invention are intended to be within the scope of the invention. For example, one may recognize that information fusion according to the present invention occurs at several levels and at several points in the information selection process. Thus, information fusion is used in generating a constraint list, and in altering confidence scores associated with speech recognition hypotheses. A multi-layered information fusion-based form filling system is thus within the scope of the present invention, and various embodiments may be realized with respect to various types of available inputs, various modalities of input, and various applications of form filling. Such variations are not to be regarded as a departure from the spirit and scope of the invention.
Claims (27)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/245,144 US20030115060A1 (en) | 2001-12-13 | 2002-09-16 | System and interactive form filling with fusion of data from multiple unreliable information sources |
US10/392,549 US7240007B2 (en) | 2001-12-13 | 2003-03-20 | Speaker authentication by fusion of voiceprint match attempt results with additional information |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/022,023 US7124085B2 (en) | 2001-12-13 | 2001-12-13 | Constraint-based speech recognition system and method |
US10/245,144 US20030115060A1 (en) | 2001-12-13 | 2002-09-16 | System and interactive form filling with fusion of data from multiple unreliable information sources |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/022,023 Continuation-In-Part US7124085B2 (en) | 2001-12-13 | 2001-12-13 | Constraint-based speech recognition system and method |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/392,549 Continuation-In-Part US7240007B2 (en) | 2001-12-13 | 2003-03-20 | Speaker authentication by fusion of voiceprint match attempt results with additional information |
Publications (1)
Publication Number | Publication Date |
---|---|
US20030115060A1 true US20030115060A1 (en) | 2003-06-19 |
Family
ID=21807419
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/022,023 Active 2024-10-30 US7124085B2 (en) | 2001-12-13 | 2001-12-13 | Constraint-based speech recognition system and method |
US10/245,144 Abandoned US20030115060A1 (en) | 2001-12-13 | 2002-09-16 | System and interactive form filling with fusion of data from multiple unreliable information sources |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/022,023 Active 2024-10-30 US7124085B2 (en) | 2001-12-13 | 2001-12-13 | Constraint-based speech recognition system and method |
Country Status (4)
Country | Link |
---|---|
US (2) | US7124085B2 (en) |
EP (1) | EP1326234A3 (en) |
JP (1) | JP2003216181A (en) |
CN (1) | CN1238832C (en) |
Cited By (138)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050288934A1 (en) * | 2004-06-29 | 2005-12-29 | Canon Kabushiki Kaisha | Multimodal input method |
US20060009974A1 (en) * | 2004-07-09 | 2006-01-12 | Matsushita Electric Industrial Co., Ltd. | Hands-free voice dialing for portable and remote devices |
US20060074652A1 (en) * | 2004-09-20 | 2006-04-06 | International Business Machines Corporation | Method and system for voice-enabled autofill |
US20060136564A1 (en) * | 2004-11-19 | 2006-06-22 | W.A. Krapf, Inc. | Bi-directional communication between a web client and a web server |
US20060167685A1 (en) * | 2002-02-07 | 2006-07-27 | Eric Thelen | Method and device for the rapid, pattern-recognition-supported transcription of spoken and written utterances |
US20060167684A1 (en) * | 2005-01-24 | 2006-07-27 | Delta Electronics, Inc. | Speech recognition method and system |
US20060276916A1 (en) * | 2004-12-22 | 2006-12-07 | Dearing Stephen M | System and method for electronically processing address information |
US20070094155A1 (en) * | 2005-05-17 | 2007-04-26 | Dearing Stephen M | System and method for automated management of an address database |
US20080281582A1 (en) * | 2007-05-11 | 2008-11-13 | Delta Electronics, Inc. | Input system for mobile search and method therefor |
US7502835B1 (en) * | 2004-11-17 | 2009-03-10 | Juniper Networks, Inc. | Virtual folders for tracking HTTP sessions |
US20090187410A1 (en) * | 2008-01-22 | 2009-07-23 | At&T Labs, Inc. | System and method of providing speech processing in user interface |
US20090248415A1 (en) * | 2008-03-31 | 2009-10-01 | Yap, Inc. | Use of metadata to post process speech recognition output |
US20100057713A1 (en) * | 2008-09-03 | 2010-03-04 | International Business Machines Corporation | Entity-driven logic for improved name-searching in mixed-entity lists |
US20100070360A1 (en) * | 2008-09-13 | 2010-03-18 | At&T Intellectual Property I, L.P. | System and method for creating a speech search platform for coupons |
US20110169906A1 (en) * | 2010-01-13 | 2011-07-14 | Seizo Suzuki | Optical scanning device and image forming apparatus |
US8024347B2 (en) | 2007-09-27 | 2011-09-20 | International Business Machines Corporation | Method and apparatus for automatically differentiating between types of names stored in a data collection |
US20110264994A1 (en) * | 2010-04-13 | 2011-10-27 | ForenSight, Inc. | Method of retrieving interactive electronic data and accessing productivity tools for real property valuation and mortgage underwriting software systems |
US20110282653A1 (en) * | 2008-12-26 | 2011-11-17 | Akihiro Tamura | Text processing apparatus, text processing method, and computer-readable recording medium |
US20120078627A1 (en) * | 2010-09-27 | 2012-03-29 | Wagner Oliver P | Electronic device with text error correction based on voice recognition data |
US20140095971A1 (en) * | 2012-09-28 | 2014-04-03 | Interactive Memories, Inc. | Method for Form Filling an Address on a Mobile Computing Device based on Zip Code Lookup |
US8892446B2 (en) | 2010-01-18 | 2014-11-18 | Apple Inc. | Service orchestration for intelligent automated assistant |
US9190062B2 (en) | 2010-02-25 | 2015-11-17 | Apple Inc. | User profiling for voice input processing |
US20160042732A1 (en) * | 2005-08-26 | 2016-02-11 | At&T Intellectual Property Ii, L.P. | System and method for robust access and entry to large structured data using voice form-filling |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US20160062971A1 (en) * | 2012-04-05 | 2016-03-03 | Mitesh L. THAKKER | Systems and methods to input or access data using remote submitting mechanism |
US9300784B2 (en) | 2013-06-13 | 2016-03-29 | Apple Inc. | System and method for emergency calls initiated by voice command |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9583107B2 (en) | 2006-04-05 | 2017-02-28 | Amazon Technologies, Inc. | Continuous speech transcription performance indication |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9697822B1 (en) | 2013-03-15 | 2017-07-04 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9973450B2 (en) | 2007-09-17 | 2018-05-15 | Amazon Technologies, Inc. | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10199051B2 (en) | 2013-02-07 | 2019-02-05 | Apple Inc. | Voice trigger for a digital assistant |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US11217255B2 (en) | 2017-05-16 | 2022-01-04 | Apple Inc. | Far-field extension for digital assistant services |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
Families Citing this family (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7444286B2 (en) | 2001-09-05 | 2008-10-28 | Roth Daniel L | Speech recognition using re-utterance recognition |
US7809574B2 (en) | 2001-09-05 | 2010-10-05 | Voice Signal Technologies Inc. | Word recognition using choice lists |
US7526431B2 (en) * | 2001-09-05 | 2009-04-28 | Voice Signal Technologies, Inc. | Speech recognition using ambiguous or phone key spelling and/or filtering |
JP3542578B2 (en) * | 2001-11-22 | 2004-07-14 | キヤノン株式会社 | Speech recognition apparatus and method, and program |
US7222072B2 (en) * | 2003-02-13 | 2007-05-22 | Sbc Properties, L.P. | Bio-phonetic multi-phrase speaker identity verification |
US20050049858A1 (en) * | 2003-08-25 | 2005-03-03 | Bellsouth Intellectual Property Corporation | Methods and systems for improving alphabetic speech recognition accuracy |
US6983244B2 (en) * | 2003-08-29 | 2006-01-03 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for improved speech recognition with supplementary information |
GB0322516D0 (en) * | 2003-09-25 | 2003-10-29 | Canon Europa Nv | Cellular mobile communication device |
JP4012143B2 (en) * | 2003-12-16 | 2007-11-21 | キヤノン株式会社 | Information processing apparatus and data input method |
US7363224B2 (en) * | 2003-12-30 | 2008-04-22 | Microsoft Corporation | Method for entering text |
WO2005076259A1 (en) * | 2004-02-10 | 2005-08-18 | Nec Corporation | Speech input system, electronic device, speech input method, and speech input program |
JP2006011641A (en) * | 2004-06-23 | 2006-01-12 | Fujitsu Ltd | Information input method and device |
EP1635328B1 (en) * | 2004-09-14 | 2018-05-30 | Swisscom AG | Speech recognition method constrained with a grammar received from a remote system. |
JP4027357B2 (en) * | 2004-10-08 | 2007-12-26 | キヤノン株式会社 | Character string input device and control method thereof |
JP4950024B2 (en) * | 2005-02-28 | 2012-06-13 | 本田技研工業株式会社 | Conversation system and conversation software |
WO2007049183A1 (en) * | 2005-10-27 | 2007-05-03 | Koninklijke Philips Electronics N.V. | Method and system for processing dictated information |
US8286071B1 (en) | 2006-06-29 | 2012-10-09 | Escription, Inc. | Insertion of standard text in transcriptions |
KR20090019198A (en) * | 2007-08-20 | 2009-02-25 | 삼성전자주식회사 | Method and apparatus for automatically completed text input using speech recognition |
KR101502003B1 (en) * | 2008-07-08 | 2015-03-12 | 엘지전자 주식회사 | Mobile terminal and method for inputting a text thereof |
US20120280915A1 (en) * | 2011-05-02 | 2012-11-08 | Nokia Corporation | Method and apparatus for facilitating interacting with a multimodal user interface |
US9257115B2 (en) | 2012-03-08 | 2016-02-09 | Facebook, Inc. | Device for extracting information from a dialog |
US9946699B1 (en) * | 2012-08-29 | 2018-04-17 | Intuit Inc. | Location-based speech recognition for preparation of electronic tax return |
KR102081925B1 (en) * | 2012-08-29 | 2020-02-26 | 엘지전자 주식회사 | display device and speech search method thereof |
US9361084B1 (en) | 2013-11-14 | 2016-06-07 | Google Inc. | Methods and systems for installing and executing applications |
US10008208B2 (en) * | 2014-09-18 | 2018-06-26 | Nuance Communications, Inc. | Method and apparatus for performing speaker recognition |
US10372804B2 (en) | 2016-05-17 | 2019-08-06 | Bruce HASSEL | Interactive audio validation/assistance system and methodologies |
CN109410923B (en) * | 2018-12-26 | 2022-06-10 | 中国联合网络通信集团有限公司 | Speech recognition method, apparatus, system and storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5530950A (en) * | 1993-07-10 | 1996-06-25 | International Business Machines Corporation | Audio data processing |
US5855000A (en) * | 1995-09-08 | 1998-12-29 | Carnegie Mellon University | Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input |
US5960399A (en) * | 1996-12-24 | 1999-09-28 | Gte Internetworking Incorporated | Client/server speech processor/recognizer |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5131045A (en) * | 1990-05-10 | 1992-07-14 | Roth Richard G | Audio-augmented data keying |
JPH07210190A (en) | 1993-12-30 | 1995-08-11 | Internatl Business Mach Corp <Ibm> | Method and system for voice recognition |
AU707122B2 (en) | 1994-10-25 | 1999-07-01 | British Telecommunications Public Limited Company | Voice-operated services |
US5963666A (en) * | 1995-08-18 | 1999-10-05 | International Business Machines Corporation | Confusion matrix mediated word prediction |
US5917890A (en) * | 1995-12-29 | 1999-06-29 | At&T Corp | Disambiguation of alphabetic characters in an automated call processing environment |
US6952799B2 (en) * | 1996-06-17 | 2005-10-04 | British Telecommunications | User interface for network browser including pre-processor for links embedded in hypermedia documents |
US5912949A (en) | 1996-11-05 | 1999-06-15 | Northern Telecom Limited | Voice-dialing system using both spoken names and initials in recognition |
US6078886A (en) * | 1997-04-14 | 2000-06-20 | At&T Corporation | System and method for providing remote automatic speech recognition services via a packet network |
US6044347A (en) * | 1997-08-05 | 2000-03-28 | Lucent Technologies Inc. | Methods and apparatus object-oriented rule-based dialogue management |
AU2789499A (en) * | 1998-02-25 | 1999-09-15 | Scansoft, Inc. | Generic run-time engine for interfacing between applications and speech engines |
FI981154A (en) | 1998-05-25 | 1999-11-26 | Nokia Mobile Phones Ltd | Voice identification procedure and apparatus |
TW501046B (en) * | 1999-06-11 | 2002-09-01 | Ind Tech Res Inst | A portable dialogue manager |
JP2001188781A (en) * | 1999-12-28 | 2001-07-10 | Sony Corp | Device and method for processing conversation and recording medium |
US7143043B1 (en) | 2000-04-26 | 2006-11-28 | Openwave Systems Inc. | Constrained keyboard disambiguation using voice recognition |
WO2002005263A1 (en) | 2000-07-07 | 2002-01-17 | Siemens Aktiengesellschaft | Method for voice input and voice recognition |
US6728348B2 (en) | 2000-11-30 | 2004-04-27 | Comverse, Inc. | System for storing voice recognizable identifiers using a limited input device such as a telephone key pad |
-
2001
- 2001-12-13 US US10/022,023 patent/US7124085B2/en active Active
-
2002
- 2002-09-16 US US10/245,144 patent/US20030115060A1/en not_active Abandoned
- 2002-11-22 EP EP02258061A patent/EP1326234A3/en not_active Withdrawn
- 2002-12-11 JP JP2002359296A patent/JP2003216181A/en active Pending
- 2002-12-12 CN CNB021559821A patent/CN1238832C/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5530950A (en) * | 1993-07-10 | 1996-06-25 | International Business Machines Corporation | Audio data processing |
US5855000A (en) * | 1995-09-08 | 1998-12-29 | Carnegie Mellon University | Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input |
US5960399A (en) * | 1996-12-24 | 1999-09-28 | Gte Internetworking Incorporated | Client/server speech processor/recognizer |
Cited By (198)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US20060167685A1 (en) * | 2002-02-07 | 2006-07-27 | Eric Thelen | Method and device for the rapid, pattern-recognition-supported transcription of spoken and written utterances |
US20050288934A1 (en) * | 2004-06-29 | 2005-12-29 | Canon Kabushiki Kaisha | Multimodal input method |
US7630901B2 (en) * | 2004-06-29 | 2009-12-08 | Canon Kabushiki Kaisha | Multimodal input method |
US20060009974A1 (en) * | 2004-07-09 | 2006-01-12 | Matsushita Electric Industrial Co., Ltd. | Hands-free voice dialing for portable and remote devices |
US20060074652A1 (en) * | 2004-09-20 | 2006-04-06 | International Business Machines Corporation | Method and system for voice-enabled autofill |
US7953597B2 (en) * | 2004-09-20 | 2011-05-31 | Nuance Communications, Inc. | Method and system for voice-enabled autofill |
US8065390B2 (en) | 2004-11-17 | 2011-11-22 | Juniper Networks, Inc. | Virtual folders for tracking HTTP sessions |
US7502835B1 (en) * | 2004-11-17 | 2009-03-10 | Juniper Networks, Inc. | Virtual folders for tracking HTTP sessions |
US20110093603A1 (en) * | 2004-11-17 | 2011-04-21 | Juniper Networks, Inc. | Virtual folders for tracking http sessions |
US7886061B1 (en) | 2004-11-17 | 2011-02-08 | Juniper Networks, Inc. | Virtual folders for tracking HTTP sessions |
US7461134B2 (en) * | 2004-11-19 | 2008-12-02 | W.A. Krapf, Inc. | Bi-directional communication between a web client and a web server |
US20060136564A1 (en) * | 2004-11-19 | 2006-06-22 | W.A. Krapf, Inc. | Bi-directional communication between a web client and a web server |
US7801925B2 (en) * | 2004-12-22 | 2010-09-21 | United States Postal Service | System and method for electronically processing address information |
US20060276916A1 (en) * | 2004-12-22 | 2006-12-07 | Dearing Stephen M | System and method for electronically processing address information |
US20060167684A1 (en) * | 2005-01-24 | 2006-07-27 | Delta Electronics, Inc. | Speech recognition method and system |
US20070094155A1 (en) * | 2005-05-17 | 2007-04-26 | Dearing Stephen M | System and method for automated management of an address database |
US8165909B2 (en) | 2005-05-17 | 2012-04-24 | The United States Postal Service | System and method for automated management of an address database |
US9824682B2 (en) * | 2005-08-26 | 2017-11-21 | Nuance Communications, Inc. | System and method for robust access and entry to large structured data using voice form-filling |
US20160042732A1 (en) * | 2005-08-26 | 2016-02-11 | At&T Intellectual Property Ii, L.P. | System and method for robust access and entry to large structured data using voice form-filling |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9583107B2 (en) | 2006-04-05 | 2017-02-28 | Amazon Technologies, Inc. | Continuous speech transcription performance indication |
US9117447B2 (en) | 2006-09-08 | 2015-08-25 | Apple Inc. | Using event alert text as input to an automated assistant |
US8942986B2 (en) | 2006-09-08 | 2015-01-27 | Apple Inc. | Determining user intent based on ontologies of domains |
US8930191B2 (en) | 2006-09-08 | 2015-01-06 | Apple Inc. | Paraphrasing of user requests and results by automated digital assistant |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US20080281582A1 (en) * | 2007-05-11 | 2008-11-13 | Delta Electronics, Inc. | Input system for mobile search and method therefor |
US9973450B2 (en) | 2007-09-17 | 2018-05-15 | Amazon Technologies, Inc. | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
US8024347B2 (en) | 2007-09-27 | 2011-09-20 | International Business Machines Corporation | Method and apparatus for automatically differentiating between types of names stored in a data collection |
US10381016B2 (en) | 2008-01-03 | 2019-08-13 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9530415B2 (en) | 2008-01-22 | 2016-12-27 | At&T Intellectual Property I, L.P. | System and method of providing speech processing in user interface |
US9177551B2 (en) * | 2008-01-22 | 2015-11-03 | At&T Intellectual Property I, L.P. | System and method of providing speech processing in user interface |
US20090187410A1 (en) * | 2008-01-22 | 2009-07-23 | At&T Labs, Inc. | System and method of providing speech processing in user interface |
US20090248415A1 (en) * | 2008-03-31 | 2009-10-01 | Yap, Inc. | Use of metadata to post process speech recognition output |
US8676577B2 (en) * | 2008-03-31 | 2014-03-18 | Canyon IP Holdings, LLC | Use of metadata to post process speech recognition output |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9865248B2 (en) | 2008-04-05 | 2018-01-09 | Apple Inc. | Intelligent text-to-speech conversion |
US10108612B2 (en) | 2008-07-31 | 2018-10-23 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US10235427B2 (en) | 2008-09-03 | 2019-03-19 | International Business Machines Corporation | Entity-driven logic for improved name-searching in mixed-entity lists |
US9411877B2 (en) | 2008-09-03 | 2016-08-09 | International Business Machines Corporation | Entity-driven logic for improved name-searching in mixed-entity lists |
US20100057713A1 (en) * | 2008-09-03 | 2010-03-04 | International Business Machines Corporation | Entity-driven logic for improved name-searching in mixed-entity lists |
US20100070360A1 (en) * | 2008-09-13 | 2010-03-18 | At&T Intellectual Property I, L.P. | System and method for creating a speech search platform for coupons |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US8886519B2 (en) * | 2008-12-26 | 2014-11-11 | Nec Corporation | Text processing apparatus, text processing method, and computer-readable recording medium |
US20110282653A1 (en) * | 2008-12-26 | 2011-11-17 | Akihiro Tamura | Text processing apparatus, text processing method, and computer-readable recording medium |
US11080012B2 (en) | 2009-06-05 | 2021-08-03 | Apple Inc. | Interface for a virtual digital assistant |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10795541B2 (en) | 2009-06-05 | 2020-10-06 | Apple Inc. | Intelligent organization of tasks items |
US10475446B2 (en) | 2009-06-05 | 2019-11-12 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US20110169906A1 (en) * | 2010-01-13 | 2011-07-14 | Seizo Suzuki | Optical scanning device and image forming apparatus |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US9548050B2 (en) | 2010-01-18 | 2017-01-17 | Apple Inc. | Intelligent automated assistant |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10706841B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Task flow identification based on user intent |
US8903716B2 (en) | 2010-01-18 | 2014-12-02 | Apple Inc. | Personalized vocabulary for digital assistant |
US8892446B2 (en) | 2010-01-18 | 2014-11-18 | Apple Inc. | Service orchestration for intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US9190062B2 (en) | 2010-02-25 | 2015-11-17 | Apple Inc. | User profiling for voice input processing |
US10049675B2 (en) | 2010-02-25 | 2018-08-14 | Apple Inc. | User profiling for voice input processing |
US20110264994A1 (en) * | 2010-04-13 | 2011-10-27 | ForenSight, Inc. | Method of retrieving interactive electronic data and accessing productivity tools for real property valuation and mortgage underwriting software systems |
US9075783B2 (en) * | 2010-09-27 | 2015-07-07 | Apple Inc. | Electronic device with text error correction based on voice recognition data |
US20120078627A1 (en) * | 2010-09-27 | 2012-03-29 | Wagner Oliver P | Electronic device with text error correction based on voice recognition data |
US8719014B2 (en) * | 2010-09-27 | 2014-05-06 | Apple Inc. | Electronic device with text error correction based on voice recognition data |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US10102359B2 (en) | 2011-03-21 | 2018-10-16 | Apple Inc. | Device access using voice authentication |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US10198417B2 (en) * | 2012-04-05 | 2019-02-05 | Mitesh L. THAKKER | Systems and methods to input or access data using remote submitting mechanism |
US20160062971A1 (en) * | 2012-04-05 | 2016-03-03 | Mitesh L. THAKKER | Systems and methods to input or access data using remote submitting mechanism |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US20140095971A1 (en) * | 2012-09-28 | 2014-04-03 | Interactive Memories, Inc. | Method for Form Filling an Address on a Mobile Computing Device based on Zip Code Lookup |
US9332137B2 (en) * | 2012-09-28 | 2016-05-03 | Interactive Memories Inc. | Method for form filling an address on a mobile computing device based on zip code lookup |
US10199051B2 (en) | 2013-02-07 | 2019-02-05 | Apple Inc. | Voice trigger for a digital assistant |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US9697822B1 (en) | 2013-03-15 | 2017-07-04 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9966060B2 (en) | 2013-06-07 | 2018-05-08 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10657961B2 (en) | 2013-06-08 | 2020-05-19 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US9300784B2 (en) | 2013-06-13 | 2016-03-29 | Apple Inc. | System and method for emergency calls initiated by voice command |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10083690B2 (en) | 2014-05-30 | 2018-09-25 | Apple Inc. | Better resolution when referencing to concepts |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10497365B2 (en) | 2014-05-30 | 2019-12-03 | Apple Inc. | Multi-command single utterance input method |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9668024B2 (en) | 2014-06-30 | 2017-05-30 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10904611B2 (en) | 2014-06-30 | 2021-01-26 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US10431204B2 (en) | 2014-09-11 | 2019-10-01 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US9986419B2 (en) | 2014-09-30 | 2018-05-29 | Apple Inc. | Social reminders |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US11556230B2 (en) | 2014-12-02 | 2023-01-17 | Apple Inc. | Data detection |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US10311871B2 (en) | 2015-03-08 | 2019-06-04 | Apple Inc. | Competing devices responding to voice triggers |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US11069347B2 (en) | 2016-06-08 | 2021-07-20 | Apple Inc. | Intelligent automated assistant for media exploration |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10553215B2 (en) | 2016-09-23 | 2020-02-04 | Apple Inc. | Intelligent automated assistant |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US11217255B2 (en) | 2017-05-16 | 2022-01-04 | Apple Inc. | Far-field extension for digital assistant services |
Also Published As
Publication number | Publication date |
---|---|
EP1326234A2 (en) | 2003-07-09 |
CN1238832C (en) | 2006-01-25 |
EP1326234A3 (en) | 2004-12-22 |
US7124085B2 (en) | 2006-10-17 |
JP2003216181A (en) | 2003-07-30 |
CN1424711A (en) | 2003-06-18 |
US20030115057A1 (en) | 2003-06-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20030115060A1 (en) | System and interactive form filling with fusion of data from multiple unreliable information sources | |
US6996531B2 (en) | Automated database assistance using a telephone for a speech based or text based multimedia communication mode | |
US7680661B2 (en) | Method and system for improved speech recognition | |
US5638425A (en) | Automated directory assistance system using word recognition and phoneme processing method | |
US7240007B2 (en) | Speaker authentication by fusion of voiceprint match attempt results with additional information | |
US6269153B1 (en) | Methods and apparatus for automatic call routing including disambiguating routing decisions | |
CA2248715C (en) | Method and apparatus for generating semantically consistent inputs to a dialog manager | |
EP2359364B1 (en) | System and method for recognizing proper names in dialog systems | |
US5404507A (en) | Apparatus and method for finding records in a database by formulating a query using equivalent terms which correspond to terms in the input query | |
US7949528B2 (en) | System and method for spelling recognition using speech and non-speech input | |
US5995956A (en) | Questionless cased-based knowledge and a method for constructing the same | |
EP1058236B1 (en) | Speech recognition based database query system | |
US20050177376A1 (en) | Recognition results postprocessor for use in voice recognition systems | |
KR20010041440A (en) | Knowledge-based strategies applied to n-best lists in automatic speech recognition systems | |
CN101405693A (en) | Personal synergic filtering of multimodal inputs | |
US20040122817A1 (en) | Systems and methods for associating system entities with descriptions | |
JP2014106982A (en) | System for providing automatically completed inquiry word, retrieval system, method for providing automatically completed inquiry word, and recording medium | |
US20020169618A1 (en) | Providing help information in a speech dialog system | |
CN110674276A (en) | Robot self-learning method, robot terminal, device and readable storage medium | |
WO2002052434A2 (en) | A system and method for improving accuracy of signal interpretation | |
US20050177373A1 (en) | Methods and apparatus for providing context and experience sensitive help in voice applications | |
US7899809B2 (en) | Method and system for querying a multimedia database from a telecommunication terminal | |
CN111930919B (en) | Enterprise online education APP voice interaction implementation method | |
US10832675B2 (en) | Speech recognition system with interactive spelling function | |
CN113806475A (en) | Information reply method and device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JUNQUA, JEAN-CLAUDE;STOIMENOV, KIRILL;KUHN, ROLAND;REEL/FRAME:013300/0049 Effective date: 20020913 |
|
AS | Assignment |
Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN Free format text: RE-RECORD TO CORRECT THE RECEIVING PARTY'S ADDRESS, PREVIOUSLY RECORDED AT REEL 013300, FRAME 0049.;ASSIGNORS:JUNQUA, JEAN-CLAUDE;STOIMENOV, KIRILL;KUHN, ROLAND;REEL/FRAME:013701/0263 Effective date: 20020913 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE |
|
AS | Assignment |
Owner name: SOVEREIGN PEAK VENTURES, LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.;REEL/FRAME:048830/0154 Effective date: 20190308 |