US20030115060A1 - System and interactive form filling with fusion of data from multiple unreliable information sources - Google Patents

System and interactive form filling with fusion of data from multiple unreliable information sources Download PDF

Info

Publication number
US20030115060A1
US20030115060A1 US10/245,144 US24514402A US2003115060A1 US 20030115060 A1 US20030115060 A1 US 20030115060A1 US 24514402 A US24514402 A US 24514402A US 2003115060 A1 US2003115060 A1 US 2003115060A1
Authority
US
United States
Prior art keywords
information
user
input
prompt
operable
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/245,144
Inventor
Jean-claude Junqua
Kirill Stoimenov
Roland Kuhn
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Holdings Corp
Sovereign Peak Ventures LLC
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US10/245,144 priority Critical patent/US20030115060A1/en
Assigned to MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD. reassignment MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JUNQUA, JEAN-CLAUDE, KUHN, ROLAND, STOIMENOV, KIRILL
Assigned to MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD. reassignment MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD. RE-RECORD TO CORRECT THE RECEIVING PARTY'S ADDRESS, PREVIOUSLY RECORDED AT REEL 013300, FRAME 0049. Assignors: JUNQUA, JEAN-CLAUDE, KUHN, ROLAND, STOIMENOV, KIRILL
Priority to US10/392,549 priority patent/US7240007B2/en
Publication of US20030115060A1 publication Critical patent/US20030115060A1/en
Assigned to SOVEREIGN PEAK VENTURES, LLC reassignment SOVEREIGN PEAK VENTURES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/24Speech recognition using non-acoustical features

Definitions

  • the present invention generally relates to speech recognition and particularly relates to automated form filling over a telephone system.
  • the information fusion process should not trust the information sources to be reliable, and should intelligently use information from the multiple sources to constrain and supplement one another based on differences in reliability between sources and of particular information inputs.
  • the fusion process should be able to incorporate a dialogue with a user to increase knowledge relating to reliability of information content, and/or gather additional information inputs.
  • an automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources.
  • An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources.
  • a form filler is operable to fill an electronic form with the selected information.
  • the form filling system of the present invention is advantageous over previous form filling systems in that it has knowledge relating to reliability of multiple information sources and is able to compare and select information content accordingly. It is further advantageous in its further aspects, wherein a prompt formulator is capable of formulating a prompt based on a comparison of the information content of the sources, and based on a level of knowledge relating to reliability of the source content, wherein the formulated prompt is designed to elicit a response from a user operable to increase knowledge of reliability relating to the multiple information sources.
  • a resulting dialog with a user succeeds in filling out an electronic form while simultaneously decreasing effort on the user's part and increasing reliability of the form contents.
  • the present invention is particularly suitable for use with electronic form filling over a telephone, wherein user input and accessible databases are unreliable information sources. Further areas of applicability of the present invention will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.
  • FIG. 1 is an illustrated block diagram depicting a general implementation of the present invention with a telephone call center
  • FIG. 2 is a flow chart depicting the method of the present invention
  • FIG. 3 is a block diagram depicting a form filling system according to the present invention.
  • FIG. 4 is a partial block and information flow diagram depicting a detailed implementation of the present invention with a telephone call center.
  • the present invention is employed to perform form filling by fusing multimodal user input over a telephone with one or more user information databases.
  • the form filling system 100 of FIG. 1 exemplifies one implementation of the preferred embodiment of the present invention to fill out a form requiring a user's name, address, and telephone number.
  • a user one Mr. Baker, is initially prompted to speak his name into the telephone receiver as at 102 , and to spell his name via the telephone keypad as at 104 .
  • the speech input 106 is communicated to an automatic speech recognizer 108
  • the keypad entry 110 is communicated to an information database 112 of names indexed by predefined classes defined by the telephone keypad.
  • a constraint list 114 of candidate names is generated from the keypad entry 110 and the information database 112 of names, and the generated constraint list 114 is further communicated to the speech recognizer 108 .
  • the speech recognizer functions to recognize the speech input 106 by generating a plurality of speech recognition hypotheses, and then selects the N best generated hypotheses by comparing them to the constraint list 114 . Thus, if “Bater” were one of the original speech recognition hypotheses, it would be discarded based on it's absence from the constraint list.
  • the information database 112 further serves as an information source having the names and addresses of most residents of the nation in which the user, Mr. Baker, is located, so frequency information 116 relating to the frequency with which names appear in the database can be communicated to rescoring module 118 used to further rescore the N best speech recognition hypotheses.
  • rescoring module 118 used to further rescore the N best speech recognition hypotheses.
  • the user is additionally prompted to enter his or her postal code, such as a zip code, and the received postal code 120 , the rescored speech recognition hypotheses, and the information database 112 are communicated to an information fuser 122 .
  • the information fuser 122 selects information content of the information database 112 based on the rescored recognition hypotheses and the received zip code 120 .
  • the information fuser 122 does not merely trust the database 112 , the rescored hypotheses, and the received zip code 120 to be accurate. Instead, the information fuser 122 recognizes and adjusts for several potential causes of unreliability.
  • the data communicated to information fuser 122 may be unreliable for several reasons.
  • the user's name may be spoken in such a way that the speech recognition system misrecognizes it.
  • the customer may mistype his or her postal code.
  • the customer database may be incomplete (address missing) or incorrect (name or address misspelled, information out of date).
  • the information fuser employs an information selection strategy that selects the most reliable information available.
  • the information fuser 122 selects the most reliable information available based on knowledge relating to reliability of various types of information sources, and based on a comparison of the information content from the different sources. For example, if the top-ranking speech recognition hypothesis does not match any name in the information database 112 having the received zip code 120 , but the second-highest ranking speech recognition hypothesis has only a slightly lower score than the top-ranking hypothesis and does match a name in the information database 112 having the received zip code 120 , then the information fuser can select the name indicated by the second-highest ranking speech recognition hypothesis and prompt the user for confirmation.
  • the information fuser 122 can select the name and address matching the highest ranked speech recognition hypothesis and/or prompt the user to reenter the postal code or confirm whether the postal code is correct. Further, the information fuser may take a different approach by distinguishing between an entirely incorrect postal code and one that is only partially incorrect, and further consider the first two digits of the postal code (in the case of a zip code) more reliable than the last three digits of the postal code.
  • Form filling system 100 preferably has an interviewer 123 for implementing a strategy to prompt the user for input increasing knowledge relating to reliability of information content.
  • the interviewer 123 can be and/or incorporate a human agent to assist in confirming, correcting, selecting, and/or supplementing information.
  • the interviewer 123 can be partially or wholly automated using a prompt formulator to request different inputs from the user in different situations.
  • information fuser 122 can select, deselect, and/or reselect information content based on the increased knowledge relating to the reliability of the information content, and the user responses constitute new information sources to be used in the fusion process.
  • the prompt formulator can preferably ask the user to supply supplemental information for fields of the form 124 for which reliable information cannot be obtained with the information sources currently available.
  • Form filling system 100 also has a form filler for filling appropriate fields of the electronic form 124 with the selected information content.
  • the method 200 of the present invention is illustrated in FIG. 2, and begins at 202 . Thence, the method 200 proceeds to step 203 , wherein information inputs are sought, preferably by initially prompting a user for specific inputs where a user is an applicable information source. Thence, the method proceeds to step 204 , wherein multiple information inputs are received from multiple information sources. In addition to information inputs provided by a user, these information inputs may also include data from an information database, and/or additional data such as that provided by caller ID or a biometric (measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.) of a user.
  • biometric measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.
  • step 206 information content is selected from one or more of the information sources based on a comparison of source contents and knowledge relating to reliability of the information sources.
  • the knowledge of reliability of a source relates to the type of source, and thus is prior knowledge.
  • the knowledge relating to reliability of an information source and/or specific information contents stems from a comparison of the contents of the information sources.
  • content of an information source of a reliable type may be deemed less reliable based on comparison with content of another information source when the information content conflicts.
  • content of an information source of an unreliable type may be deemed more reliable based on comparison with content of another information source when the information content matches. This process can be used to identify more and less reliable portions of information content within one or more information sources.
  • step 208 an electronic form is filled with the selected information. If the form is deemed reliably completed as at 210 , then the method ends at 214 and a filled form has been generated. On the other hand, some or all of the selected information may be deemed insufficiently reliable, and/or a sufficient amount of fields of the form may not be deemed completed. In either of these latter cases, method 200 returns to step 203 .
  • step 203 the user is prompted for additional information inputs, and the prompt is designed to elicit a response to increase knowledge of reliability relating to selectable information and/or gather supplemental information.
  • the request for additional sources may be a request for confirmation of selected information, a request to reenter one of the user inputs, a request for a different information type, or a similar type of request.
  • speech generation is preferred to communicate the request, especially over a telephone system.
  • the prompt for additional information inputs the method 200 returns to step 204 , wherein the response is received. The method then proceeds again to step 206 , wherein new information content is selected and/or it is determined whether selected content is reliable based on the new information sources.
  • the method 200 further proceeds to step 208 and fills the form based on the revised selection. If the newly filled form is deemed reliably completed at step 210 , then the method 200 ends at 216 . Otherwise, processing continues in a recursive fashion until a reliably completed form is obtained or the process is otherwise interrupted.
  • a form filling system 300 of the present invention is more generally illustrated in FIG. 3, wherein a first information input 302 from a first information source and a second information input 304 from a second information source are received by input 306 and communicated to information fuser 308 .
  • Information fuser 308 compares and selects information based on reliability of the information as determined based on the comparison and in accordance with predefined rules developed with regard to reliability of different types of information sources and/or information inputs.
  • the selected information 310 is communicated to form filler 312 , which fills an electronic form 314 with the selected information to generate a filled form 316 and communicate it to an outside system via output 316 .
  • Information fuser 308 mutually communicates with prompt formulator 320 , and prompt formulator 320 formulates a prompt 322 for additional information sources, if needed, based on reliability of the information sources, comparative and/or objective reliability of the available information inputs, and/or requirements for supplemental information content.
  • Prompt formulator 320 further communicates its current state to information fuser 308 so that information fuser 308 is aware of the type of information input(s) requested and how to interpret its information content in view of the other information inputs.
  • the formulated prompt 322 is communicated to a dialogue manager 324 that generates a prompt in a manner communicable to and understandable by a user, preferably by speech generation.
  • the generated prompt is communicated to the user via output 328 .
  • a response from the user constitutes an additional information input communicable to information fuser 308 via input 306 .
  • multi-modal information inputs 400 from an information source corresponding to a user include a user speech input 402 and a user keypad entry 404 .
  • the user speech input includes a spoken user name
  • the user keypad entry includes a spelling and/or initials of the user name and a zip code for the user location.
  • textual information inputs 406 from database information sources include nickname data input 408 from a nickname database and personal data input 410 from a personal information database. Use of nickname data input 408 in the present detailed implementation represents an improvement over the more simplified implementation previously discussed with reference to FIG. 1.
  • additional data inputs 412 that also represent an improvement include telephone subscriber data 414 provided by a caller ID service.
  • Other examples of additional data inputs 412 that can also be used include biometrics identifying a user and gathered, for example, by a handheld device the user employs to communicate with the call center. Further examples include an IP address for the user location, assuming the user communicates with the call center using a computer network.
  • Form filling system 100 receives the multi-modal information inputs 400 , textual information inputs 406 , and additional data inputs 412 and uses the various information inputs to constrain and supplement one another according to their varying modalities, utilities, and levels of reliability.
  • user speech input 402 is processed by speech recognizer 108 A to produce a plurality of speech recognition hypotheses 414 .
  • user keypad entry 404 , additional data inputs 412 , nickname data input 408 , and personal data input 410 are all communicated to constraint list generator 416 .
  • Constraint list generator 416 is adapted in the present implementation to generate a constraint list of candidate names by using a user keypad entry 404 containing a first and last name is used to access the nickname database and generate a plurality of first names based on the input first name, such that a keypad entry generating a constraint list containing “Bob” will also contain “Robert”, “Robby”, “Bobby”, and so on.
  • the plurality of first names thus generated, a zip code from the user keypad entry, and any matching caller ID information are then used to access the personal information database, such as Phonedisc, containing names and addresses of all telephone subscribes in the United States, to generate a constraint list 114 that is communicated to N best hypotheses generator 108 B.
  • the speech recognition hypotheses 414 are processes by N best hypotheses generator 108 B to generate an N best list of speech recognition hypotheses 418 .
  • N best list of speech recognition hypotheses 418 is communicated to rescoring mechanism 118 , as is data from the personal information database pertaining to frequency of appearance of names in the personal information database.
  • Rescoring mechanism 118 rescores the N best list of speech recognition hypotheses 418 to generate a list of rescored hypotheses 420 .
  • Caller ID information and a postal code portion of the user keypad entry 404 may alternatively and/or additionally be used during this process to rescore hypotheses based on the frequency of names in the database having the corresponding postal code and/or names matching the caller ID information.
  • the rescored hypotheses 420 , caller ID information, postal code, and personal data input 410 are communicated to the information fuser 122 and the prompt formulator 320 .
  • the information fuser selects most reliable information from the personal data input 410 , caller ID information, and rescored hypotheses 420 , and the selected data 310 is communicated to the prompt formulator 320 and to form filler 312 .
  • Form filler fills electronic form 314 to produce a filled form 316 that is partially or wholly filled with the selected data.
  • the prompt formulator 320 determines whether more information is needed based on the selected data, the information available for selection, and requirements of the electronic form. If more information inputs are required, the prompt formulator formulates an appropriate prompt 322 and communicates a current state 422 to information fuser 122 . Otherwise, the form is deemed completed.

Abstract

An automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources. An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources. A form filler is operable to fill an electronic form with the selected information.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation-in-part of U.S. patent application Ser. No. 10/022,023 filed on Dec. 13, 2001. The disclosure of the above application is incorporated herein by reference.[0001]
  • FIELD OF THE INVENTION
  • The present invention generally relates to speech recognition and particularly relates to automated form filling over a telephone system. [0002]
  • BACKGROUND OF THE INVENTION
  • Automatic electronic form filling by a user, particularly over a telephone system, is a notoriously laborious and error prone process. Use of numerical keypad entries to attempt retrieval of user information is one existing process that proves to be only as reliable as completeness and correctness of the user information database and user knowledge of the required information and successful operation of the telephone keypad. Other applications, such as confirming availability and/or delivery of a product over the phone or on a networked computer system, suffer from similar problems. [0003]
  • The key to improving the automated form filling process, by increasing reliability of information and decreasing effort on the part of a user, involves recognizing that information from multiple information sources can be fused in an intelligent manner. To be successful, the information fusion process should not trust the information sources to be reliable, and should intelligently use information from the multiple sources to constrain and supplement one another based on differences in reliability between sources and of particular information inputs. To be further successful, the fusion process should be able to incorporate a dialogue with a user to increase knowledge relating to reliability of information content, and/or gather additional information inputs. [0004]
  • Automated form filling processes do not currently succeed in intelligently fusing information from multiple information sources based on knowledge relating to differences in reliability information form different sources, thereby simultaneously increasing reliability of form contents and decreasing effort on the part of the user. The need remains, therefore, for a solution to the problems associated with automated form filling as detailed above. The present invention provides such a solution. [0005]
  • SUMMARY OF THE INVENTION
  • In accordance with the present invention, an automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources. An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources. A form filler is operable to fill an electronic form with the selected information. [0006]
  • The form filling system of the present invention is advantageous over previous form filling systems in that it has knowledge relating to reliability of multiple information sources and is able to compare and select information content accordingly. It is further advantageous in its further aspects, wherein a prompt formulator is capable of formulating a prompt based on a comparison of the information content of the sources, and based on a level of knowledge relating to reliability of the source content, wherein the formulated prompt is designed to elicit a response from a user operable to increase knowledge of reliability relating to the multiple information sources. A resulting dialog with a user succeeds in filling out an electronic form while simultaneously decreasing effort on the user's part and increasing reliability of the form contents. [0007]
  • The present invention is particularly suitable for use with electronic form filling over a telephone, wherein user input and accessible databases are unreliable information sources. Further areas of applicability of the present invention will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.[0008]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention will become more fully understood from the detailed description and the accompanying drawings, wherein: [0009]
  • FIG. 1 is an illustrated block diagram depicting a general implementation of the present invention with a telephone call center; [0010]
  • FIG. 2 is a flow chart depicting the method of the present invention; [0011]
  • FIG. 3 is a block diagram depicting a form filling system according to the present invention; and [0012]
  • FIG. 4 is a partial block and information flow diagram depicting a detailed implementation of the present invention with a telephone call center.[0013]
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • The following description of the preferred embodiment(s) is merely exemplary in nature and is in no way intended to limit the invention, its application, or uses. In a preferred embodiment, the present invention is employed to perform form filling by fusing multimodal user input over a telephone with one or more user information databases. The [0014] form filling system 100 of FIG. 1 exemplifies one implementation of the preferred embodiment of the present invention to fill out a form requiring a user's name, address, and telephone number.
  • According to form [0015] filling system 100, a user, one Mr. Baker, is initially prompted to speak his name into the telephone receiver as at 102, and to spell his name via the telephone keypad as at 104. The speech input 106 is communicated to an automatic speech recognizer 108, whereas the keypad entry 110 is communicated to an information database 112 of names indexed by predefined classes defined by the telephone keypad. In turn, a constraint list 114 of candidate names is generated from the keypad entry 110 and the information database 112 of names, and the generated constraint list 114 is further communicated to the speech recognizer 108. The speech recognizer functions to recognize the speech input 106 by generating a plurality of speech recognition hypotheses, and then selects the N best generated hypotheses by comparing them to the constraint list 114. Thus, if “Bater” were one of the original speech recognition hypotheses, it would be discarded based on it's absence from the constraint list.
  • The [0016] information database 112 further serves as an information source having the names and addresses of most residents of the nation in which the user, Mr. Baker, is located, so frequency information 116 relating to the frequency with which names appear in the database can be communicated to rescoring module 118 used to further rescore the N best speech recognition hypotheses. Thus, if “Baker” and “Bakes” are both present in the N best speech recognition hypotheses, they can be rescored to increase the ranking of “Baker” with respect to the ranking of “Bakes” based on a higher frequency of “Baker” in the information database 112 compared to the frequency of “Bakes”.
  • The user is additionally prompted to enter his or her postal code, such as a zip code, and the received [0017] postal code 120, the rescored speech recognition hypotheses, and the information database 112 are communicated to an information fuser 122. The information fuser 122, in turn, selects information content of the information database 112 based on the rescored recognition hypotheses and the received zip code 120. The information fuser 122, however, does not merely trust the database 112, the rescored hypotheses, and the received zip code 120 to be accurate. Instead, the information fuser 122 recognizes and adjusts for several potential causes of unreliability.
  • The data communicated to [0018] information fuser 122 may be unreliable for several reasons. For example, the user's name may be spoken in such a way that the speech recognition system misrecognizes it. Also, the customer may mistype his or her postal code. Further, the customer database may be incomplete (address missing) or incorrect (name or address misspelled, information out of date). For these and similar reasons, the information fuser employs an information selection strategy that selects the most reliable information available.
  • The information fuser [0019] 122 selects the most reliable information available based on knowledge relating to reliability of various types of information sources, and based on a comparison of the information content from the different sources. For example, if the top-ranking speech recognition hypothesis does not match any name in the information database 112 having the received zip code 120, but the second-highest ranking speech recognition hypothesis has only a slightly lower score than the top-ranking hypothesis and does match a name in the information database 112 having the received zip code 120, then the information fuser can select the name indicated by the second-highest ranking speech recognition hypothesis and prompt the user for confirmation. Similarly, if the top-ranking speech recognition hypothesis does not match any name in the information database 112 having the received postal code 120, but has a much higher score than the second-highest ranking speech recognition hypothesis, then the information fuser 122 can select the name and address matching the highest ranked speech recognition hypothesis and/or prompt the user to reenter the postal code or confirm whether the postal code is correct. Further, the information fuser may take a different approach by distinguishing between an entirely incorrect postal code and one that is only partially incorrect, and further consider the first two digits of the postal code (in the case of a zip code) more reliable than the last three digits of the postal code.
  • [0020] Form filling system 100 preferably has an interviewer 123 for implementing a strategy to prompt the user for input increasing knowledge relating to reliability of information content. The interviewer 123 can be and/or incorporate a human agent to assist in confirming, correcting, selecting, and/or supplementing information. Alternatively, the interviewer 123 can be partially or wholly automated using a prompt formulator to request different inputs from the user in different situations. Accordingly, information fuser 122 can select, deselect, and/or reselect information content based on the increased knowledge relating to the reliability of the information content, and the user responses constitute new information sources to be used in the fusion process. Thus, the prompt formulator can preferably ask the user to supply supplemental information for fields of the form 124 for which reliable information cannot be obtained with the information sources currently available. Form filling system 100 also has a form filler for filling appropriate fields of the electronic form 124 with the selected information content.
  • The [0021] method 200 of the present invention is illustrated in FIG. 2, and begins at 202. Thence, the method 200 proceeds to step 203, wherein information inputs are sought, preferably by initially prompting a user for specific inputs where a user is an applicable information source. Thence, the method proceeds to step 204, wherein multiple information inputs are received from multiple information sources. In addition to information inputs provided by a user, these information inputs may also include data from an information database, and/or additional data such as that provided by caller ID or a biometric (measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.) of a user. Thence, the method 200 proceeds to step 206, wherein information content is selected from one or more of the information sources based on a comparison of source contents and knowledge relating to reliability of the information sources. In one aspect, the knowledge of reliability of a source relates to the type of source, and thus is prior knowledge. In another aspect, however, the knowledge relating to reliability of an information source and/or specific information contents stems from a comparison of the contents of the information sources. Thus, content of an information source of a reliable type may be deemed less reliable based on comparison with content of another information source when the information content conflicts. Similarly, content of an information source of an unreliable type may be deemed more reliable based on comparison with content of another information source when the information content matches. This process can be used to identify more and less reliable portions of information content within one or more information sources.
  • Once selection of information has taken place, [0022] method 200 proceeds to step 208, wherein an electronic form is filled with the selected information. If the form is deemed reliably completed as at 210, then the method ends at 214 and a filled form has been generated. On the other hand, some or all of the selected information may be deemed insufficiently reliable, and/or a sufficient amount of fields of the form may not be deemed completed. In either of these latter cases, method 200 returns to step 203.
  • In [0023] step 203, the user is prompted for additional information inputs, and the prompt is designed to elicit a response to increase knowledge of reliability relating to selectable information and/or gather supplemental information. Thus, the request for additional sources may be a request for confirmation of selected information, a request to reenter one of the user inputs, a request for a different information type, or a similar type of request. With this step, speech generation is preferred to communicate the request, especially over a telephone system. Following, the prompt for additional information inputs, the method 200 returns to step 204, wherein the response is received. The method then proceeds again to step 206, wherein new information content is selected and/or it is determined whether selected content is reliable based on the new information sources. The method 200 further proceeds to step 208 and fills the form based on the revised selection. If the newly filled form is deemed reliably completed at step 210, then the method 200 ends at 216. Otherwise, processing continues in a recursive fashion until a reliably completed form is obtained or the process is otherwise interrupted.
  • A [0024] form filling system 300 of the present invention is more generally illustrated in FIG. 3, wherein a first information input 302 from a first information source and a second information input 304 from a second information source are received by input 306 and communicated to information fuser 308. Information fuser 308 compares and selects information based on reliability of the information as determined based on the comparison and in accordance with predefined rules developed with regard to reliability of different types of information sources and/or information inputs. The selected information 310 is communicated to form filler 312, which fills an electronic form 314 with the selected information to generate a filled form 316 and communicate it to an outside system via output 316.
  • [0025] Information fuser 308 mutually communicates with prompt formulator 320, and prompt formulator 320 formulates a prompt 322 for additional information sources, if needed, based on reliability of the information sources, comparative and/or objective reliability of the available information inputs, and/or requirements for supplemental information content. Prompt formulator 320 further communicates its current state to information fuser 308 so that information fuser 308 is aware of the type of information input(s) requested and how to interpret its information content in view of the other information inputs. The formulated prompt 322 is communicated to a dialogue manager 324 that generates a prompt in a manner communicable to and understandable by a user, preferably by speech generation. The generated prompt is communicated to the user via output 328. A response from the user constitutes an additional information input communicable to information fuser 308 via input 306.
  • A detailed implementation of the present invention with a telephone call center is described with reference to FIG. 4. Therein, [0026] multi-modal information inputs 400 from an information source corresponding to a user include a user speech input 402 and a user keypad entry 404. The user speech input includes a spoken user name, and the user keypad entry includes a spelling and/or initials of the user name and a zip code for the user location. Also, textual information inputs 406 from database information sources include nickname data input 408 from a nickname database and personal data input 410 from a personal information database. Use of nickname data input 408 in the present detailed implementation represents an improvement over the more simplified implementation previously discussed with reference to FIG. 1. Further, additional data inputs 412 that also represent an improvement include telephone subscriber data 414 provided by a caller ID service. Other examples of additional data inputs 412 that can also be used include biometrics identifying a user and gathered, for example, by a handheld device the user employs to communicate with the call center. Further examples include an IP address for the user location, assuming the user communicates with the call center using a computer network. These types of information inputs all assist to varying degrees in identifying the caller, and the form filling system of the present invention is adapted to use some or all of these types of information inputs accordingly.
  • [0027] Form filling system 100 receives the multi-modal information inputs 400, textual information inputs 406, and additional data inputs 412 and uses the various information inputs to constrain and supplement one another according to their varying modalities, utilities, and levels of reliability. For example, user speech input 402 is processed by speech recognizer 108A to produce a plurality of speech recognition hypotheses 414. Also, user keypad entry 404, additional data inputs 412, nickname data input 408, and personal data input 410 are all communicated to constraint list generator 416.
  • [0028] Constraint list generator 416 is adapted in the present implementation to generate a constraint list of candidate names by using a user keypad entry 404 containing a first and last name is used to access the nickname database and generate a plurality of first names based on the input first name, such that a keypad entry generating a constraint list containing “Bob” will also contain “Robert”, “Robby”, “Bobby”, and so on. The plurality of first names thus generated, a zip code from the user keypad entry, and any matching caller ID information are then used to access the personal information database, such as Phonedisc, containing names and addresses of all telephone subscribes in the United States, to generate a constraint list 114 that is communicated to N best hypotheses generator 108B.
  • The [0029] speech recognition hypotheses 414 are processes by N best hypotheses generator 108B to generate an N best list of speech recognition hypotheses 418. One skilled in the art of speech recognition will recognize that an alternative embodiment may be realized with a word lattice. The N best list of speech recognition hypotheses 418 is communicated to rescoring mechanism 118, as is data from the personal information database pertaining to frequency of appearance of names in the personal information database. Rescoring mechanism 118 rescores the N best list of speech recognition hypotheses 418 to generate a list of rescored hypotheses 420. Caller ID information and a postal code portion of the user keypad entry 404 may alternatively and/or additionally be used during this process to rescore hypotheses based on the frequency of names in the database having the corresponding postal code and/or names matching the caller ID information.
  • The rescored [0030] hypotheses 420, caller ID information, postal code, and personal data input 410 are communicated to the information fuser 122 and the prompt formulator 320. The information fuser selects most reliable information from the personal data input 410, caller ID information, and rescored hypotheses 420, and the selected data 310 is communicated to the prompt formulator 320 and to form filler 312. Form filler fills electronic form 314 to produce a filled form 316 that is partially or wholly filled with the selected data. Meanwhile, the prompt formulator 320 determines whether more information is needed based on the selected data, the information available for selection, and requirements of the electronic form. If more information inputs are required, the prompt formulator formulates an appropriate prompt 322 and communicates a current state 422 to information fuser 122. Otherwise, the form is deemed completed.
  • The description of the invention is merely exemplary in nature and, thus, variations that do not depart from the gist of the invention are intended to be within the scope of the invention. For example, one may recognize that information fusion according to the present invention occurs at several levels and at several points in the information selection process. Thus, information fusion is used in generating a constraint list, and in altering confidence scores associated with speech recognition hypotheses. A multi-layered information fusion-based form filling system is thus within the scope of the present invention, and various embodiments may be realized with respect to various types of available inputs, various modalities of input, and various applications of form filling. Such variations are not to be regarded as a departure from the spirit and scope of the invention. [0031]

Claims (27)

What is claimed is:
1. An automated form filling system comprising:
an input receptive of a plurality of information inputs from a plurality of information sources, including a first information input from a first information source and a second information input from a second information source;
an information fuser operable to select information from at least one of the first information input and the second information input based on a comparison of the first information input and the second information input, and based on knowledge relating to reliability of the first information source and the second information source; and
a form filler operable to fill an electronic form with the selected information.
2. The system of claim 1, wherein said information fuser is operable to select information from at least one of the first information input and the second information input based on knowledge relating to reliability of the first information input and the second information input.
3. The system of claim 1 comprising a prompt formulator operable to formulate a prompt designed to elicit a response from a user operable to increase knowledge relating to reliability of selectable information.
4. The system of claim 3 comprising a dialog manager operable to receive the formulated prompt and communicate the formulated prompt to a user in a manner so as to be readily understood by the user.
5. The system of claim 4, wherein said dialog manager operable to generate speech based on the formulated prompt and to communicate the generated speech to the user.
6. The system of claim 3, wherein said prompt formulator includes a human performing a review of at least one of the selected information and the plurality of information inputs, formulating the prompt based on the review, and at least one of confirming and correcting the selected information entered in the filled form based on the elicited responses.
7. The system of claim 3, wherein said prompt formulator is operable to formulate the prompt based on a comparison of the first information input and the second information input, and based on knowledge relating to reliability of the first information source and the second information source.
8. The system of claim 1, wherein the first information source corresponds to the user, and the second information source corresponds to an information database.
9. The system of claim 8, wherein the first information input corresponds to at least one of a user speech input and a user keypad entry.
10. The system of claim 9, wherein the first information input corresponds to the user speech input, wherein said plurality of information inputs includes a third information input from the user, wherein the third information input corresponds to a user keypad entry.
11. The system of claim 1, wherein said information fuser is adapted to yield numerical scores for various portions of information of the second information input based on a comparison of the first information input and the second information input.
12. The system of claim 1, wherein said information fuser is adapted to yield numerical scores for various portions of information of the second information input based on knowledge relating to reliability of various portions of information of the second information input.
13. The system of claim 1, wherein the first information source corresponds to a user, and the first information input corresponds to a measured physical characteristic of the user.
14. The system of claim 1, wherein the first information source corresponds to a user providing the first information input via telephone at a user location, and wherein the second information source corresponds to an automated caller identification system providing at least one of a phone number at the user location and a name associated with the user location.
15. The system of claim 1, wherein the first information source corresponds to a user, and wherein the first information input corresponds to a user handwritten entry.
16. The system of claim 1, wherein the first information source corresponds to a user, the first information input corresponds to a name of a user, the second information source corresponds to a database of names and other personal data of potential users, and the information fuser is operable to make a comparison of the name of the user and the names of potential users, and operable to select at least one of a name and associated personal data from the database based on the comparison.
17. A data fusion-based method of automated form filling comprising:
receiving a plurality of information inputs from a plurality of information sources;
selecting information from the plurality of information inputs based on a comparison of the information inputs and knowledge relating to reliability of the information sources; and
filling an electronic form with the selected information.
18. The method of claim 17, wherein said selecting information from the plurality of information inputs is based on knowledge relating to reliability of the information inputs.
19. The method of claim 17, wherein said receiving a plurality of information inputs from a plurality of information sources corresponds to receiving a user input from a user.
20. The method of claim 19 comprising:
formulating a prompt designed to elicit a new information input from the user; and
prompting the user for the new information input.
21. The method of claim 19, wherein said formulating the prompt includes formulating a prompt designed to elicit a new information input from the user that increases knowledge relating to reliability of selectable information.
22. The method of claim 19, wherein said formulating the prompt includes formulating a prompt designed to elicit a new information input from the user that supplements selectable information.
23. The method of claim 19 comprising selecting unselected information based on the new information input.
24. The method of claim 19 comprising deselecting selected information based on the new information input.
25. The method of claim 19 comprising comparing the new information input to the information inputs previously received.
26. A data fusion-based form filling system for filling an electronic form based on user input communicated over a telephone system, comprising:
an input receptive of personal information input by a user via a telephone system using input speech and a keypad entry;
a speech recognizer operable to recognize the personal information based on the input speech, the keypad entry, and knowledge relating to reliability of the input speech and the keypad entry;
an electronic form filler operable to access a database of stored personal information for potential users, operable to make a comparison between the recognized personal information and the stored personal information, operable to select personal information for the user based on the comparison and based on knowledge relating to reliability of the recognized personal information and the stored personal information, and operable to fill an electronic form with the selected personal information.
27. The system of claim 26 comprising an interviewer operable to prompt the user for additional input required to assist in at least one of:
(a) selecting the personal information;
(b) confirming the selected personal information;
(c) correcting the selected personal information; and
(d) supplementing the selected personal information.
US10/245,144 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources Abandoned US20030115060A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US10/245,144 US20030115060A1 (en) 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources
US10/392,549 US7240007B2 (en) 2001-12-13 2003-03-20 Speaker authentication by fusion of voiceprint match attempt results with additional information

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/022,023 US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method
US10/245,144 US20030115060A1 (en) 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/022,023 Continuation-In-Part US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US10/392,549 Continuation-In-Part US7240007B2 (en) 2001-12-13 2003-03-20 Speaker authentication by fusion of voiceprint match attempt results with additional information

Publications (1)

Publication Number Publication Date
US20030115060A1 true US20030115060A1 (en) 2003-06-19

Family

ID=21807419

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/022,023 Active 2024-10-30 US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method
US10/245,144 Abandoned US20030115060A1 (en) 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10/022,023 Active 2024-10-30 US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method

Country Status (4)

Country Link
US (2) US7124085B2 (en)
EP (1) EP1326234A3 (en)
JP (1) JP2003216181A (en)
CN (1) CN1238832C (en)

Cited By (138)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050288934A1 (en) * 2004-06-29 2005-12-29 Canon Kabushiki Kaisha Multimodal input method
US20060009974A1 (en) * 2004-07-09 2006-01-12 Matsushita Electric Industrial Co., Ltd. Hands-free voice dialing for portable and remote devices
US20060074652A1 (en) * 2004-09-20 2006-04-06 International Business Machines Corporation Method and system for voice-enabled autofill
US20060136564A1 (en) * 2004-11-19 2006-06-22 W.A. Krapf, Inc. Bi-directional communication between a web client and a web server
US20060167685A1 (en) * 2002-02-07 2006-07-27 Eric Thelen Method and device for the rapid, pattern-recognition-supported transcription of spoken and written utterances
US20060167684A1 (en) * 2005-01-24 2006-07-27 Delta Electronics, Inc. Speech recognition method and system
US20060276916A1 (en) * 2004-12-22 2006-12-07 Dearing Stephen M System and method for electronically processing address information
US20070094155A1 (en) * 2005-05-17 2007-04-26 Dearing Stephen M System and method for automated management of an address database
US20080281582A1 (en) * 2007-05-11 2008-11-13 Delta Electronics, Inc. Input system for mobile search and method therefor
US7502835B1 (en) * 2004-11-17 2009-03-10 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US20090187410A1 (en) * 2008-01-22 2009-07-23 At&T Labs, Inc. System and method of providing speech processing in user interface
US20090248415A1 (en) * 2008-03-31 2009-10-01 Yap, Inc. Use of metadata to post process speech recognition output
US20100057713A1 (en) * 2008-09-03 2010-03-04 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US20100070360A1 (en) * 2008-09-13 2010-03-18 At&T Intellectual Property I, L.P. System and method for creating a speech search platform for coupons
US20110169906A1 (en) * 2010-01-13 2011-07-14 Seizo Suzuki Optical scanning device and image forming apparatus
US8024347B2 (en) 2007-09-27 2011-09-20 International Business Machines Corporation Method and apparatus for automatically differentiating between types of names stored in a data collection
US20110264994A1 (en) * 2010-04-13 2011-10-27 ForenSight, Inc. Method of retrieving interactive electronic data and accessing productivity tools for real property valuation and mortgage underwriting software systems
US20110282653A1 (en) * 2008-12-26 2011-11-17 Akihiro Tamura Text processing apparatus, text processing method, and computer-readable recording medium
US20120078627A1 (en) * 2010-09-27 2012-03-29 Wagner Oliver P Electronic device with text error correction based on voice recognition data
US20140095971A1 (en) * 2012-09-28 2014-04-03 Interactive Memories, Inc. Method for Form Filling an Address on a Mobile Computing Device based on Zip Code Lookup
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9190062B2 (en) 2010-02-25 2015-11-17 Apple Inc. User profiling for voice input processing
US20160042732A1 (en) * 2005-08-26 2016-02-11 At&T Intellectual Property Ii, L.P. System and method for robust access and entry to large structured data using voice form-filling
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US20160062971A1 (en) * 2012-04-05 2016-03-03 Mitesh L. THAKKER Systems and methods to input or access data using remote submitting mechanism
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9583107B2 (en) 2006-04-05 2017-02-28 Amazon Technologies, Inc. Continuous speech transcription performance indication
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9973450B2 (en) 2007-09-17 2018-05-15 Amazon Technologies, Inc. Methods and systems for dynamically updating web service profile information by parsing transcribed message strings
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7444286B2 (en) 2001-09-05 2008-10-28 Roth Daniel L Speech recognition using re-utterance recognition
US7809574B2 (en) 2001-09-05 2010-10-05 Voice Signal Technologies Inc. Word recognition using choice lists
US7526431B2 (en) * 2001-09-05 2009-04-28 Voice Signal Technologies, Inc. Speech recognition using ambiguous or phone key spelling and/or filtering
JP3542578B2 (en) * 2001-11-22 2004-07-14 キヤノン株式会社 Speech recognition apparatus and method, and program
US7222072B2 (en) * 2003-02-13 2007-05-22 Sbc Properties, L.P. Bio-phonetic multi-phrase speaker identity verification
US20050049858A1 (en) * 2003-08-25 2005-03-03 Bellsouth Intellectual Property Corporation Methods and systems for improving alphabetic speech recognition accuracy
US6983244B2 (en) * 2003-08-29 2006-01-03 Matsushita Electric Industrial Co., Ltd. Method and apparatus for improved speech recognition with supplementary information
GB0322516D0 (en) * 2003-09-25 2003-10-29 Canon Europa Nv Cellular mobile communication device
JP4012143B2 (en) * 2003-12-16 2007-11-21 キヤノン株式会社 Information processing apparatus and data input method
US7363224B2 (en) * 2003-12-30 2008-04-22 Microsoft Corporation Method for entering text
WO2005076259A1 (en) * 2004-02-10 2005-08-18 Nec Corporation Speech input system, electronic device, speech input method, and speech input program
JP2006011641A (en) * 2004-06-23 2006-01-12 Fujitsu Ltd Information input method and device
EP1635328B1 (en) * 2004-09-14 2018-05-30 Swisscom AG Speech recognition method constrained with a grammar received from a remote system.
JP4027357B2 (en) * 2004-10-08 2007-12-26 キヤノン株式会社 Character string input device and control method thereof
JP4950024B2 (en) * 2005-02-28 2012-06-13 本田技研工業株式会社 Conversation system and conversation software
WO2007049183A1 (en) * 2005-10-27 2007-05-03 Koninklijke Philips Electronics N.V. Method and system for processing dictated information
US8286071B1 (en) 2006-06-29 2012-10-09 Escription, Inc. Insertion of standard text in transcriptions
KR20090019198A (en) * 2007-08-20 2009-02-25 삼성전자주식회사 Method and apparatus for automatically completed text input using speech recognition
KR101502003B1 (en) * 2008-07-08 2015-03-12 엘지전자 주식회사 Mobile terminal and method for inputting a text thereof
US20120280915A1 (en) * 2011-05-02 2012-11-08 Nokia Corporation Method and apparatus for facilitating interacting with a multimodal user interface
US9257115B2 (en) 2012-03-08 2016-02-09 Facebook, Inc. Device for extracting information from a dialog
US9946699B1 (en) * 2012-08-29 2018-04-17 Intuit Inc. Location-based speech recognition for preparation of electronic tax return
KR102081925B1 (en) * 2012-08-29 2020-02-26 엘지전자 주식회사 display device and speech search method thereof
US9361084B1 (en) 2013-11-14 2016-06-07 Google Inc. Methods and systems for installing and executing applications
US10008208B2 (en) * 2014-09-18 2018-06-26 Nuance Communications, Inc. Method and apparatus for performing speaker recognition
US10372804B2 (en) 2016-05-17 2019-08-06 Bruce HASSEL Interactive audio validation/assistance system and methodologies
CN109410923B (en) * 2018-12-26 2022-06-10 中国联合网络通信集团有限公司 Speech recognition method, apparatus, system and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5530950A (en) * 1993-07-10 1996-06-25 International Business Machines Corporation Audio data processing
US5855000A (en) * 1995-09-08 1998-12-29 Carnegie Mellon University Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input
US5960399A (en) * 1996-12-24 1999-09-28 Gte Internetworking Incorporated Client/server speech processor/recognizer

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5131045A (en) * 1990-05-10 1992-07-14 Roth Richard G Audio-augmented data keying
JPH07210190A (en) 1993-12-30 1995-08-11 Internatl Business Mach Corp <Ibm> Method and system for voice recognition
AU707122B2 (en) 1994-10-25 1999-07-01 British Telecommunications Public Limited Company Voice-operated services
US5963666A (en) * 1995-08-18 1999-10-05 International Business Machines Corporation Confusion matrix mediated word prediction
US5917890A (en) * 1995-12-29 1999-06-29 At&T Corp Disambiguation of alphabetic characters in an automated call processing environment
US6952799B2 (en) * 1996-06-17 2005-10-04 British Telecommunications User interface for network browser including pre-processor for links embedded in hypermedia documents
US5912949A (en) 1996-11-05 1999-06-15 Northern Telecom Limited Voice-dialing system using both spoken names and initials in recognition
US6078886A (en) * 1997-04-14 2000-06-20 At&T Corporation System and method for providing remote automatic speech recognition services via a packet network
US6044347A (en) * 1997-08-05 2000-03-28 Lucent Technologies Inc. Methods and apparatus object-oriented rule-based dialogue management
AU2789499A (en) * 1998-02-25 1999-09-15 Scansoft, Inc. Generic run-time engine for interfacing between applications and speech engines
FI981154A (en) 1998-05-25 1999-11-26 Nokia Mobile Phones Ltd Voice identification procedure and apparatus
TW501046B (en) * 1999-06-11 2002-09-01 Ind Tech Res Inst A portable dialogue manager
JP2001188781A (en) * 1999-12-28 2001-07-10 Sony Corp Device and method for processing conversation and recording medium
US7143043B1 (en) 2000-04-26 2006-11-28 Openwave Systems Inc. Constrained keyboard disambiguation using voice recognition
WO2002005263A1 (en) 2000-07-07 2002-01-17 Siemens Aktiengesellschaft Method for voice input and voice recognition
US6728348B2 (en) 2000-11-30 2004-04-27 Comverse, Inc. System for storing voice recognizable identifiers using a limited input device such as a telephone key pad

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5530950A (en) * 1993-07-10 1996-06-25 International Business Machines Corporation Audio data processing
US5855000A (en) * 1995-09-08 1998-12-29 Carnegie Mellon University Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input
US5960399A (en) * 1996-12-24 1999-09-28 Gte Internetworking Incorporated Client/server speech processor/recognizer

Cited By (198)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US20060167685A1 (en) * 2002-02-07 2006-07-27 Eric Thelen Method and device for the rapid, pattern-recognition-supported transcription of spoken and written utterances
US20050288934A1 (en) * 2004-06-29 2005-12-29 Canon Kabushiki Kaisha Multimodal input method
US7630901B2 (en) * 2004-06-29 2009-12-08 Canon Kabushiki Kaisha Multimodal input method
US20060009974A1 (en) * 2004-07-09 2006-01-12 Matsushita Electric Industrial Co., Ltd. Hands-free voice dialing for portable and remote devices
US20060074652A1 (en) * 2004-09-20 2006-04-06 International Business Machines Corporation Method and system for voice-enabled autofill
US7953597B2 (en) * 2004-09-20 2011-05-31 Nuance Communications, Inc. Method and system for voice-enabled autofill
US8065390B2 (en) 2004-11-17 2011-11-22 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US7502835B1 (en) * 2004-11-17 2009-03-10 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US20110093603A1 (en) * 2004-11-17 2011-04-21 Juniper Networks, Inc. Virtual folders for tracking http sessions
US7886061B1 (en) 2004-11-17 2011-02-08 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US7461134B2 (en) * 2004-11-19 2008-12-02 W.A. Krapf, Inc. Bi-directional communication between a web client and a web server
US20060136564A1 (en) * 2004-11-19 2006-06-22 W.A. Krapf, Inc. Bi-directional communication between a web client and a web server
US7801925B2 (en) * 2004-12-22 2010-09-21 United States Postal Service System and method for electronically processing address information
US20060276916A1 (en) * 2004-12-22 2006-12-07 Dearing Stephen M System and method for electronically processing address information
US20060167684A1 (en) * 2005-01-24 2006-07-27 Delta Electronics, Inc. Speech recognition method and system
US20070094155A1 (en) * 2005-05-17 2007-04-26 Dearing Stephen M System and method for automated management of an address database
US8165909B2 (en) 2005-05-17 2012-04-24 The United States Postal Service System and method for automated management of an address database
US9824682B2 (en) * 2005-08-26 2017-11-21 Nuance Communications, Inc. System and method for robust access and entry to large structured data using voice form-filling
US20160042732A1 (en) * 2005-08-26 2016-02-11 At&T Intellectual Property Ii, L.P. System and method for robust access and entry to large structured data using voice form-filling
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9583107B2 (en) 2006-04-05 2017-02-28 Amazon Technologies, Inc. Continuous speech transcription performance indication
US9117447B2 (en) 2006-09-08 2015-08-25 Apple Inc. Using event alert text as input to an automated assistant
US8942986B2 (en) 2006-09-08 2015-01-27 Apple Inc. Determining user intent based on ontologies of domains
US8930191B2 (en) 2006-09-08 2015-01-06 Apple Inc. Paraphrasing of user requests and results by automated digital assistant
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US20080281582A1 (en) * 2007-05-11 2008-11-13 Delta Electronics, Inc. Input system for mobile search and method therefor
US9973450B2 (en) 2007-09-17 2018-05-15 Amazon Technologies, Inc. Methods and systems for dynamically updating web service profile information by parsing transcribed message strings
US8024347B2 (en) 2007-09-27 2011-09-20 International Business Machines Corporation Method and apparatus for automatically differentiating between types of names stored in a data collection
US10381016B2 (en) 2008-01-03 2019-08-13 Apple Inc. Methods and apparatus for altering audio output signals
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9530415B2 (en) 2008-01-22 2016-12-27 At&T Intellectual Property I, L.P. System and method of providing speech processing in user interface
US9177551B2 (en) * 2008-01-22 2015-11-03 At&T Intellectual Property I, L.P. System and method of providing speech processing in user interface
US20090187410A1 (en) * 2008-01-22 2009-07-23 At&T Labs, Inc. System and method of providing speech processing in user interface
US20090248415A1 (en) * 2008-03-31 2009-10-01 Yap, Inc. Use of metadata to post process speech recognition output
US8676577B2 (en) * 2008-03-31 2014-03-18 Canyon IP Holdings, LLC Use of metadata to post process speech recognition output
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US10108612B2 (en) 2008-07-31 2018-10-23 Apple Inc. Mobile device having human language translation capability with positional feedback
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US10235427B2 (en) 2008-09-03 2019-03-19 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US9411877B2 (en) 2008-09-03 2016-08-09 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US20100057713A1 (en) * 2008-09-03 2010-03-04 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US20100070360A1 (en) * 2008-09-13 2010-03-18 At&T Intellectual Property I, L.P. System and method for creating a speech search platform for coupons
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US8886519B2 (en) * 2008-12-26 2014-11-11 Nec Corporation Text processing apparatus, text processing method, and computer-readable recording medium
US20110282653A1 (en) * 2008-12-26 2011-11-17 Akihiro Tamura Text processing apparatus, text processing method, and computer-readable recording medium
US11080012B2 (en) 2009-06-05 2021-08-03 Apple Inc. Interface for a virtual digital assistant
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10795541B2 (en) 2009-06-05 2020-10-06 Apple Inc. Intelligent organization of tasks items
US10475446B2 (en) 2009-06-05 2019-11-12 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US20110169906A1 (en) * 2010-01-13 2011-07-14 Seizo Suzuki Optical scanning device and image forming apparatus
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10706841B2 (en) 2010-01-18 2020-07-07 Apple Inc. Task flow identification based on user intent
US8903716B2 (en) 2010-01-18 2014-12-02 Apple Inc. Personalized vocabulary for digital assistant
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US9190062B2 (en) 2010-02-25 2015-11-17 Apple Inc. User profiling for voice input processing
US10049675B2 (en) 2010-02-25 2018-08-14 Apple Inc. User profiling for voice input processing
US20110264994A1 (en) * 2010-04-13 2011-10-27 ForenSight, Inc. Method of retrieving interactive electronic data and accessing productivity tools for real property valuation and mortgage underwriting software systems
US9075783B2 (en) * 2010-09-27 2015-07-07 Apple Inc. Electronic device with text error correction based on voice recognition data
US20120078627A1 (en) * 2010-09-27 2012-03-29 Wagner Oliver P Electronic device with text error correction based on voice recognition data
US8719014B2 (en) * 2010-09-27 2014-05-06 Apple Inc. Electronic device with text error correction based on voice recognition data
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US10102359B2 (en) 2011-03-21 2018-10-16 Apple Inc. Device access using voice authentication
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US10198417B2 (en) * 2012-04-05 2019-02-05 Mitesh L. THAKKER Systems and methods to input or access data using remote submitting mechanism
US20160062971A1 (en) * 2012-04-05 2016-03-03 Mitesh L. THAKKER Systems and methods to input or access data using remote submitting mechanism
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US20140095971A1 (en) * 2012-09-28 2014-04-03 Interactive Memories, Inc. Method for Form Filling an Address on a Mobile Computing Device based on Zip Code Lookup
US9332137B2 (en) * 2012-09-28 2016-05-03 Interactive Memories Inc. Method for form filling an address on a mobile computing device based on zip code lookup
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10978090B2 (en) 2013-02-07 2021-04-13 Apple Inc. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US11388291B2 (en) 2013-03-14 2022-07-12 Apple Inc. System and method for processing voicemail
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10657961B2 (en) 2013-06-08 2020-05-19 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US10169329B2 (en) 2014-05-30 2019-01-01 Apple Inc. Exemplar-based natural language processing
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10083690B2 (en) 2014-05-30 2018-09-25 Apple Inc. Better resolution when referencing to concepts
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US11257504B2 (en) 2014-05-30 2022-02-22 Apple Inc. Intelligent assistant for home automation
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US10497365B2 (en) 2014-05-30 2019-12-03 Apple Inc. Multi-command single utterance input method
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10904611B2 (en) 2014-06-30 2021-01-26 Apple Inc. Intelligent automated assistant for TV user interactions
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US10431204B2 (en) 2014-09-11 2019-10-01 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US11556230B2 (en) 2014-12-02 2023-01-17 Apple Inc. Data detection
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US11087759B2 (en) 2015-03-08 2021-08-10 Apple Inc. Virtual assistant activation
US10311871B2 (en) 2015-03-08 2019-06-04 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US11069347B2 (en) 2016-06-08 2021-07-20 Apple Inc. Intelligent automated assistant for media exploration
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US11152002B2 (en) 2016-06-11 2021-10-19 Apple Inc. Application integration with a digital assistant
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10553215B2 (en) 2016-09-23 2020-02-04 Apple Inc. Intelligent automated assistant
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services

Also Published As

Publication number Publication date
EP1326234A2 (en) 2003-07-09
CN1238832C (en) 2006-01-25
EP1326234A3 (en) 2004-12-22
US7124085B2 (en) 2006-10-17
JP2003216181A (en) 2003-07-30
CN1424711A (en) 2003-06-18
US20030115057A1 (en) 2003-06-19

Similar Documents

Publication Publication Date Title
US20030115060A1 (en) System and interactive form filling with fusion of data from multiple unreliable information sources
US6996531B2 (en) Automated database assistance using a telephone for a speech based or text based multimedia communication mode
US7680661B2 (en) Method and system for improved speech recognition
US5638425A (en) Automated directory assistance system using word recognition and phoneme processing method
US7240007B2 (en) Speaker authentication by fusion of voiceprint match attempt results with additional information
US6269153B1 (en) Methods and apparatus for automatic call routing including disambiguating routing decisions
CA2248715C (en) Method and apparatus for generating semantically consistent inputs to a dialog manager
EP2359364B1 (en) System and method for recognizing proper names in dialog systems
US5404507A (en) Apparatus and method for finding records in a database by formulating a query using equivalent terms which correspond to terms in the input query
US7949528B2 (en) System and method for spelling recognition using speech and non-speech input
US5995956A (en) Questionless cased-based knowledge and a method for constructing the same
EP1058236B1 (en) Speech recognition based database query system
US20050177376A1 (en) Recognition results postprocessor for use in voice recognition systems
KR20010041440A (en) Knowledge-based strategies applied to n-best lists in automatic speech recognition systems
CN101405693A (en) Personal synergic filtering of multimodal inputs
US20040122817A1 (en) Systems and methods for associating system entities with descriptions
JP2014106982A (en) System for providing automatically completed inquiry word, retrieval system, method for providing automatically completed inquiry word, and recording medium
US20020169618A1 (en) Providing help information in a speech dialog system
CN110674276A (en) Robot self-learning method, robot terminal, device and readable storage medium
WO2002052434A2 (en) A system and method for improving accuracy of signal interpretation
US20050177373A1 (en) Methods and apparatus for providing context and experience sensitive help in voice applications
US7899809B2 (en) Method and system for querying a multimedia database from a telecommunication terminal
CN111930919B (en) Enterprise online education APP voice interaction implementation method
US10832675B2 (en) Speech recognition system with interactive spelling function
CN113806475A (en) Information reply method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JUNQUA, JEAN-CLAUDE;STOIMENOV, KIRILL;KUHN, ROLAND;REEL/FRAME:013300/0049

Effective date: 20020913

AS Assignment

Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN

Free format text: RE-RECORD TO CORRECT THE RECEIVING PARTY'S ADDRESS, PREVIOUSLY RECORDED AT REEL 013300, FRAME 0049.;ASSIGNORS:JUNQUA, JEAN-CLAUDE;STOIMENOV, KIRILL;KUHN, ROLAND;REEL/FRAME:013701/0263

Effective date: 20020913

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE

AS Assignment

Owner name: SOVEREIGN PEAK VENTURES, LLC, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.;REEL/FRAME:048830/0154

Effective date: 20190308