WO2009120449A1 - Intra-language statistical machine translation - Google Patents

Intra-language statistical machine translation Download PDF

Info

Publication number
WO2009120449A1
WO2009120449A1 PCT/US2009/035389 US2009035389W WO2009120449A1 WO 2009120449 A1 WO2009120449 A1 WO 2009120449A1 US 2009035389 W US2009035389 W US 2009035389W WO 2009120449 A1 WO2009120449 A1 WO 2009120449A1
Authority
WO
WIPO (PCT)
Prior art keywords
language
query
translation model
model
statistical
Prior art date
Application number
PCT/US2009/035389
Other languages
French (fr)
Inventor
Xiao Li
Yun-Cheng Ju
Geoffrey Zweig
Alex Acero
Original Assignee
Microsoft Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Corporation filed Critical Microsoft Corporation
Priority to EP09725780.2A priority Critical patent/EP2269148B1/en
Priority to CN200980112180.XA priority patent/CN101981566B/en
Publication of WO2009120449A1 publication Critical patent/WO2009120449A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/42Data-driven translation
    • G06F40/44Statistical methods, e.g. probability models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/42Data-driven translation
    • G06F40/45Example-based machine translation; Alignment

Definitions

  • Voice search involves a coupling of voice recognition and information retrieval. An uttered phrase is automatically recognized as text, and the text is submitted as a query to a search service. For example, a person may use a mobile phone equipped with a voice search application to find a restaurant by speaking the name of the restaurant into the mobile device, and the mobile device may recognize the spoken restaurant name (i.e., convert it to text) and transmit the text of the restaurant name to a remote search service such as a business directory.
  • Local search is a special case of search where listings of business establishments, firms, organizations, or other entities have been used to enable mobile devices to search same. Consider the following example.
  • a user may be interested in finding information about a business listed in a directory as "Kung Ho Cuisine of China". However, the user formulates a query as "Kung Ho Restaurant". Currently, a search for this listing will not take advantage of statistical parallels between parts of the query and listing forms. Furthermore, erroneous listings, e.g. "Kung Ho Grocery" may be returned as a relevant match.
  • Training data may be provided.
  • the training data may include pairs of source phrases and target phrases.
  • the pairs may be used to train an intra-language statistical machine translation model, where the intra-language statistical machine translation model, when given an input phrase of text in the human language, can compute probabilities of semantic equivalence of the input phrase to possible translations of the input phrase in the human language.
  • the statistical machine translation model may be used to translate between queries and listings.
  • the queries may be text strings in the human language submitted to a search engine.
  • the listing strings may be text strings of formal names of real world entities that are to be searched by the search engine to find matches for the query strings.
  • Figure 1 shows a general process for intra-language statistical machine translation.
  • Figure 2 shows a process for building an n-gram based model.
  • Figure 3 shows an arrangement for using a statistical translation model to improve a search system and/or the language model of a voice recognition system.
  • a statistical translation model is a generalization of some sample of text, which may be parallel phrases such as a query strings and corresponding directory listings. Some types of statistical translation models give probabilities that a target sentence or phrase is a translation of a source sentence or phrase, and the probabilities reflect the statistical patterns derived from the training text. In effect, the model is a probabilistic generalization of characteristics or trends reflected from statistical measurements of training sentences.
  • Figure 1 shows a general process for intra-language statistical machine translation. Initially, a statistical machine translation model is trained 1 00. Training 1 00 will be described in detail later.
  • the training 1 00 is performed using a sample of training data, which may come from a variety of sources.
  • the training data will include parallel (paired) phrases in a same human language.
  • the training 1 00 informs the translation model with statistics (e.g., n-grams) that can be used to compute probabilities or likelihoods of candidate translations of a phrase. Specific training 1 00 for an n-gram based model will be described below.
  • the model is used to translate 1 02 a source phrase into a target phrase.
  • Translation 1 02 involves starting with a source phrase and obtaining a semantically similar or equivalent target phrase. For example, a source phrase "Kung Ho Cuisine of China" might be translated to a target phrase "Kung Ho Chinese Restaurant” or "Kung Ho Restaurant” Different forms of candidate target phrases are obtained.
  • the statistical translation model is used to find one or more of the most probable candidate target phrases.
  • a voice search system may involve two components: a voice recognition component and an information retrieval (search) component.
  • Statistical LMs e.g. n-gram models, are often used to allow flexibility in what users can say. That is, they allow a variety of sayings to be recognized by the ASR component.
  • documents d may be in the form of business listings (names of business, organizations, or other entities), which are typically short, e.g. "Kung Ho Cuisine of China”.
  • listings and queries because they are both relatively short, are treated as pairs akin to "sentence pairs" found in bilingual translation training.
  • a bilingual statistical translation model adapted for intra-language translation, may be used to automatically convert the original form of a listing to its query forms (i.e., the forms that a user might be expected to input when searching for the listing), which in turn may be used for building more robust LMs for voice search, grammar checking, or other applications.
  • the statistical translation model may be trained using a small number of transcribed or artificially produced queries, without necessarily having to acquire matching listings. While a variety of types of statistical models can be used for machine translation, an n-gram based model will be described next.
  • a query phrase and its intended listing phrase may differ in form, there is usually a semantic correspondence, at the word level, between the two phrases.
  • words in the query can be mapped to words in the listing or to a null word, and vice versa.
  • a machine translation approach may be used to predict possible query forms of a listing, and then to utilize the predicted query forms to improve language modeling.
  • n-grams may be used on word pairs to model the joint (conditional) probability of a listing and a query.
  • Figure 2 shows a process for building an n-gram based model.
  • a pair of source and target sentences in a same human language are received 1 20.
  • An alignment between the source and target sentences is obtained 1 22 by computing an edit distance between the two sentences.
  • Words and/or phrases of the aligned sentences are then paired 1 24 and treated as semantic units. Pairings may be formed by finding semantically/literally similar/equivalent words or phrases.
  • the pairings are then used to train 1 26 an n-gram model. The steps of this process may be repeated for different source and target sentences. While a small set of training sentences may suffice for some applications, using more training data will create a more robust model.
  • initial training data is provided.
  • This data may be a body of parallel text (d, q), where listings d and queries q serve as source and target sentences respectively.
  • the sentences d and q may be monotonically aligned, where null words are added, if necessary, to account for insertions or deletions that occur in the alignment.
  • the monotonic alignment will be denoted as a. Note that in another embodiment, a non-monotonic alignment may be used.
  • a listing-to-query translation may be performed. Given a listing form d, and given query forms q (from a decoder, discussed later), the query forms are searched to find those that have the highest conditional probability: q * « max q max ⁇ p M (d, q, ⁇ ) (4) where p(d, q, a) is evaluated using equation (3).
  • the translation not only exploits word-level semantic correspondence as modeled by unigrams, but it also takes into account word context by using higher order n- grams.
  • the search for the best or n-best query forms can be achieved efficiently by applying the best-first search algorithm, which is described by Russell and Norvig in Artificial Intelligence: A Modern Approach (Prentice Hall, second edition, 2003). Using this type of search, pruning techniques may be applied to reduce computational complexity.
  • LM language model
  • the n-best query forms are obtained for the listings, they may be used as training sentences for LM estimation.
  • null words in d raises a potential problem at decode time — the search space is significantly expanded because null can be present or absent at any position of the source sentence.
  • it is preferable to eliminate the use of (d, null, q,) as semantic units for values of q,.
  • q.-iq,) or (d, + i , q,q ⁇ + i) may be treated as a single semantic unit.
  • null is not explicitly inserted in d, because using semantic units (d,-i, q,-i q,) or (d, + i, q,q ⁇ + i) is equivalent to adding null in the source sentence.
  • OOV out-of-vocabulary
  • a re-ordering strategy may be used. This may be implemented before a monotonic alignment is applied by reordering d while keeping the order of q.
  • the best way to reorder the words in the source form is determined by computing the resulting joint n-gram model likelihood. Only orders that are shifts of the original order are considered, and a maximum entropy classifier for these orders is built, where the input of the classifier is the source form, and the output is an order. Prior to translation, this classifier is applied to reorder a source form.
  • Figure 3 shows an arrangement for using a statistical translation model to improve a search system and/or the language model of a voice recognition system.
  • a search engine 1 52 is configured to search listings 1 54, for example business listings.
  • the search engine 1 52 receives text queries or transcribed spoken queries 1 56 that are generated by users and submitted to the search engine 1 52.
  • Corresponding relevant listings 1 58 are retrieved by the search engine 1 52.
  • training pairs can also be obtained algorithmically using TF-IDF (term frequency-inverse document frequency).
  • the text or transcribed queries 1 56 and corresponding search-engine retrieved listings 1 58 are passed to a training component 1 60 that trains a statistical translation model 1 62, which may be n-gram based or another type of model.
  • a statistical translation model 1 62 which may be n-gram based or another type of model.
  • the training component 1 60 iterates through source-target pairs of the transcribed queries 1 56 and listings 1 58.
  • an n-gram based model given a (source, target) pair, an initial monotonic alignment is obtained between the source form and target form by computing an edit distance. Given the alignment, the training component 1 60 discovers word-level pairs and builds an n-gram translation model 1 62 based on the word- level pairs.
  • the alignment and n-gram model parameters of the translation model 1 62 may be iteratively refined to improve the translation model 1 62. Furthermore, training may implement a backoff strategy which assumes that a word can be translated to itself, as is possible with intra-language translation. In other words, the aligned units WORD-WORD, where WORD can be a word or a phrase, will have a positive probability.
  • a translation module 1 64 uses the translation model 1 62 to test decoded candidates (potential translations). Given the trained translation model 1 62 and a source form, a best-first search algorithm may be used to obtain the top n-best target forms (the n decoded target forms with the highest probability according to the translation model 1 62).
  • the weight of each target form is determined by p(target
  • d) s(q, x) ⁇ . Alternatively, relevancy may be measured directly from the translation probability, in which case s(q, d) p(q,d). In one embodiment, potential translations can be filtered out if their similarity measure is below a specified threshold.
  • a language model 1 68 can also be built or augmented using intra-language translation.
  • Language models are used in many natural language processing applications such as ASR, machine translation, and parsing.
  • the intra-language translation provided by the translation model 1 62 and translation module 1 64 may be used in language modeling by translating listings into query forms and using the same-language translated query forms when estimating a language model 1 68.
  • the count of a translated query form may be set to its posterior probabilities multiplied by the count of its original listing.
  • a server- or client-based voice recognizer may be provided with the language model 1 68, which will allow the voice recognizer to perform more accurate and comprehensive speech recognition with respect to utterances directed to the listings 1 54 or to listings.
  • the translation model 1 62 may also be used at a server or at a mobile client to translate a string inputted at the mobile device (whether by ASR or otherwise) to a display form.
  • Embodiments and features discussed above can be realized in the form of information stored in volatile or non-volatile computer or device readable media. This is deemed to include at least media such as optical storage (e.g., CD-ROM), magnetic media, flash ROM, or any current or future means of storing digital information.
  • the stored information can be in the form of machine executable instructions (e.g., compiled executable binary code), source code, bytecode, or any other information that can be used to enable or configure computing devices to perform the various embodiments discussed above.
  • This is also deemed to include at least volatile memory such as RAM and/or virtual memory storing information such as CPU instructions during execution of a program carrying out an embodiment, as well as non-volatile media storing information that allows a program or executable to be loaded and executed.
  • volatile memory such as RAM and/or virtual memory storing information such as CPU instructions during execution of a program carrying out an embodiment
  • non-volatile media storing information that allows a program or executable to be loaded and executed.
  • the embodiments and featured can be performed on any type of computing device, including portable devices, workstations, servers, mobile wireless devices, and so on.
  • the modules, components, processes, and search engine 1 52 discussed above may by realized on one computing device or multiple cooperating computing devices.

Abstract

Training data may be provided, the training data including pairs of source phrases and target phrases. The pairs may be used to train an intra-language statistical machine translation model, where the intra-language statistical machine translation model, when given an input phrase of text in the human language, can compute probabilities of semantic equivalence of the input phrase to possible translations of the input phrase in the human language. The statistical machine translation model may be used to translate between queries and listings. The queries may be text strings in the human language submitted to a search engine. The listing strings may be text strings of formal names of real world entities that are to be searched by the search engine to find matches for the query strings.

Description

INTRA-LANGUAGE STATISTICAL MACHINE TRANSLATION
BACKGROUND [0001 ] Network based search services, Internet search engines, voice search, local search, and various other technologies for searching and retrieving information have become increasingly important for helping people find information. Voice search involves a coupling of voice recognition and information retrieval. An uttered phrase is automatically recognized as text, and the text is submitted as a query to a search service. For example, a person may use a mobile phone equipped with a voice search application to find a restaurant by speaking the name of the restaurant into the mobile device, and the mobile device may recognize the spoken restaurant name (i.e., convert it to text) and transmit the text of the restaurant name to a remote search service such as a business directory. Local search is a special case of search where listings of business establishments, firms, organizations, or other entities have been used to enable mobile devices to search same. Consider the following example.
[0002] A user may be interested in finding information about a business listed in a directory as "Kung Ho Cuisine of China". However, the user formulates a query as "Kung Ho Restaurant". Currently, a search for this listing will not take advantage of statistical parallels between parts of the query and listing forms. Furthermore, erroneous listings, e.g. "Kung Ho Grocery" may be returned as a relevant match.
[0003] Discussed below are techniques related to statistical intra-language machine translation, and applications thereof to speech recognition, search, and other technologies.
SUMMARY
[0004] The following summary is included only to introduce some concepts discussed in the Detailed Description below. This summary is not comprehensive and is not intended to delineate the scope of the claimed subject matter, which is set forth by the claims presented at the end. [0005] Training data may be provided. The training data may include pairs of source phrases and target phrases. The pairs may be used to train an intra-language statistical machine translation model, where the intra-language statistical machine translation model, when given an input phrase of text in the human language, can compute probabilities of semantic equivalence of the input phrase to possible translations of the input phrase in the human language. The statistical machine translation model may be used to translate between queries and listings. The queries may be text strings in the human language submitted to a search engine. The listing strings may be text strings of formal names of real world entities that are to be searched by the search engine to find matches for the query strings.
[0006] Many of the attendant features will be explained below with reference to the following detailed description considered in connection with the accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS [0007] The present description will be better understood from the following detailed description read in light of the accompanying drawings, wherein like reference numerals are used to designate like parts in the accompanying description.
[0008] Figure 1 shows a general process for intra-language statistical machine translation. [0009] Figure 2 shows a process for building an n-gram based model.
[001 0] Figure 3 shows an arrangement for using a statistical translation model to improve a search system and/or the language model of a voice recognition system.
DETAILED DESCRIPTION OVERVIEW
[001 1 ] The description below covers embodiments related to using a statistical machine translation model to translate between sentences or phrases of a same human language. The description begins with discussion of how a relatively small set of training sentences or phrases are used to train a statistical translation model. Applications of the intra-language machine translation model are then described, including applications to search, automatic speech recognition (ASR), and display of speech recognition results
INTRALANGUAGE STATISTICAL MACHINE TRANSLATION MODEL [001 2] Statistical models have been used to translate sentences from one language to another language. However, they have not been trained or used for translating between phrases or sentences of a same language. That is, statistical modeling has not previously been used to translate phrases in English, for example, to other semantically similar phrases also in English. [001 3] A statistical translation model is a generalization of some sample of text, which may be parallel phrases such as a query strings and corresponding directory listings. Some types of statistical translation models give probabilities that a target sentence or phrase is a translation of a source sentence or phrase, and the probabilities reflect the statistical patterns derived from the training text. In effect, the model is a probabilistic generalization of characteristics or trends reflected from statistical measurements of training sentences. Note that throughout this description, the terms "sentence" and "phrase" will be used interchangeably to refer to relatively short arrangements of words. Formal and informal names of businesses, query strings inputted by users, grammatical sentences, clauses, and the like are examples of sentences or phrases. Note also that while this description discusses intra-language statistical machine translation as applied to phrase-based search (in particular voice and/or geographically localized search), the concepts are not limited to search applications. Furthermore, searching listings of short phrases is also applicable to other types of search besides local search, including product search, job search, etc. [001 4] Figure 1 shows a general process for intra-language statistical machine translation. Initially, a statistical machine translation model is trained 1 00. Training 1 00 will be described in detail later. The training 1 00 is performed using a sample of training data, which may come from a variety of sources. The training data will include parallel (paired) phrases in a same human language. The training 1 00 informs the translation model with statistics (e.g., n-grams) that can be used to compute probabilities or likelihoods of candidate translations of a phrase. Specific training 1 00 for an n-gram based model will be described below.
[001 5] After the model is trained 1 00, the model is used to translate 1 02 a source phrase into a target phrase. Translation 1 02 involves starting with a source phrase and obtaining a semantically similar or equivalent target phrase. For example, a source phrase "Kung Ho Cuisine of China" might be translated to a target phrase "Kung Ho Chinese Restaurant" or "Kung Ho Restaurant" Different forms of candidate target phrases are obtained. The statistical translation model is used to find one or more of the most probable candidate target phrases. Consider the following overview of voice based search and how it relates to intra-language machine translation.
[0016] A voice search system may involve two components: a voice recognition component and an information retrieval (search) component. A spoken utterance o is converted into a text query q using automatic speech recognition (ASR), i.e., q* = argmaxq p(o\q)p(q) (1 ) where p(o|q) and p(q) represent an acoustic model and a language model (LM), respectively. Statistical LMs, e.g. n-gram models, are often used to allow flexibility in what users can say. That is, they allow a variety of sayings to be recognized by the ASR component. Next, the best (or n-best) q is passed to a search engine to retrieve the most relevant document d, i.e. d* = argmaxd p(d\q) (2)
[001 7] In the context of local search, documents d may be in the form of business listings (names of business, organizations, or other entities), which are typically short, e.g. "Kung Ho Cuisine of China". [001 8] Given this framework for voice based search, listings and queries, because they are both relatively short, are treated as pairs akin to "sentence pairs" found in bilingual translation training. A bilingual statistical translation model, adapted for intra-language translation, may be used to automatically convert the original form of a listing to its query forms (i.e., the forms that a user might be expected to input when searching for the listing), which in turn may be used for building more robust LMs for voice search, grammar checking, or other applications. Conveniently, the statistical translation model may be trained using a small number of transcribed or artificially produced queries, without necessarily having to acquire matching listings. While a variety of types of statistical models can be used for machine translation, an n-gram based model will be described next.
[001 9] Although a query phrase and its intended listing phrase may differ in form, there is usually a semantic correspondence, at the word level, between the two phrases. In other words, words in the query can be mapped to words in the listing or to a null word, and vice versa. A machine translation approach may be used to predict possible query forms of a listing, and then to utilize the predicted query forms to improve language modeling. Specifically, as discussed next, n-grams may be used on word pairs to model the joint (conditional) probability of a listing and a query.
[0020] Figure 2 shows a process for building an n-gram based model. A pair of source and target sentences in a same human language are received 1 20. An alignment between the source and target sentences is obtained 1 22 by computing an edit distance between the two sentences. Words and/or phrases of the aligned sentences are then paired 1 24 and treated as semantic units. Pairings may be formed by finding semantically/literally similar/equivalent words or phrases. The pairings are then used to train 1 26 an n-gram model. The steps of this process may be repeated for different source and target sentences. While a small set of training sentences may suffice for some applications, using more training data will create a more robust model. Note also that the alignment and the n-gram model may be iteratively updated and refined in the maximum likelihood sense. [0021 ] Details of generating an n-gram based model will now be described. For training 1 00 an n-gram based model, initial training data is provided. This data may be a body of parallel text (d, q), where listings d and queries q serve as source and target sentences respectively. The sentences d and q may be monotonically aligned, where null words are added, if necessary, to account for insertions or deletions that occur in the alignment. The monotonic alignment will be denoted as a. Note that in another embodiment, a non-monotonic alignment may be used. [0022] Once aligned, a sequence pairs of words from d and q is generated, which is denoted as (d, q, a) = ((di , qi), (d2, q2), ..., (dι_, qθ), where each (d,, q,) is treated as a single semantic unit. Consecutive word pairs can be merged to form phrase pairs if necessary. [0023] The sequence of word pairs can then be used to train an n-gram model. Consequently, the probability of an aligned sentence pair is computed as
{d, q, a) =
1 P((rfi, '7i) l (rfi-n+l- '7ι-n+l)- - - (rfι-l- '7ι-l)) where M denotes the monotonic condition. Note that the initial alignment a may be computed using the Levenshtein distance between d and q. The alignment and the n-gram model's parameters may be updated in the maximum likelihood sense. Re-alignment can be based on pairing frequencies, for example.
[0024] Given the trained n-gram model, a listing-to-query translation may be performed. Given a listing form d, and given query forms q (from a decoder, discussed later), the query forms are searched to find those that have the highest conditional probability: q* « maxq maxα pM(d, q, α) (4) where p(d, q, a) is evaluated using equation (3).
[0025] The translation not only exploits word-level semantic correspondence as modeled by unigrams, but it also takes into account word context by using higher order n- grams. The search for the best or n-best query forms can be achieved efficiently by applying the best-first search algorithm, which is described by Russell and Norvig in Artificial Intelligence: A Modern Approach (Prentice Hall, second edition, 2003). Using this type of search, pruning techniques may be applied to reduce computational complexity. Returning to the language model (LM) for speech recognition, once the n-best query forms are obtained for the listings, they may be used as training sentences for LM estimation. [0026] There are two implementation details to be considered. First, allowing the use of null words in d raises a potential problem at decode time — the search space is significantly expanded because null can be present or absent at any position of the source sentence. To avoid this problem, it is preferable to eliminate the use of (d, = null, q,) as semantic units for values of q,. Specifically, in training, (d, = null, q,) may be merged with its proceeding or following semantic unit, depending on which of the phrases, q,_i q, or q,ql+i , have more occurrences in the training data. Then, (d.+i. q.-iq,) or (d,+i , q,qι+i) may be treated as a single semantic unit. At decode time, null is not explicitly inserted in d, because using semantic units (d,-i, q,-i q,) or (d,+i, q,qι+i) is equivalent to adding null in the source sentence. [0027] The second implementation detail concerns out-of-vocabulary (OOV) words in d. When OOV occurs, it might not be possible to produce any query forms, since p(d, = OOV, q,) = 0 for any value of q,. To deal with such cases, a positive probability may be assigned to unigrams (d, , q, = d,) whenever d, = OOV. This implies that a listing word, if never seen in the training data, will be translated to itself.
[0028] It should be noted that embodiments with non-monotonic alignment are also possible. Furthermore, a re-ordering strategy may be used. This may be implemented before a monotonic alignment is applied by reordering d while keeping the order of q. When training the translation model, the best way to reorder the words in the source form is determined by computing the resulting joint n-gram model likelihood. Only orders that are shifts of the original order are considered, and a maximum entropy classifier for these orders is built, where the input of the classifier is the source form, and the output is an order. Prior to translation, this classifier is applied to reorder a source form.
APPLICATIONS OF INTRALANGUAGE STATISTICAL TRANSLATION MODEL [0029] Figure 3 shows an arrangement for using a statistical translation model to improve a search system and/or the language model of a voice recognition system. A search engine 1 52 is configured to search listings 1 54, for example business listings. The search engine 1 52 receives text queries or transcribed spoken queries 1 56 that are generated by users and submitted to the search engine 1 52. Corresponding relevant listings 1 58 are retrieved by the search engine 1 52. Note that training pairs can also be obtained algorithmically using TF-IDF (term frequency-inverse document frequency). [0030] The text or transcribed queries 1 56 and corresponding search-engine retrieved listings 1 58 are passed to a training component 1 60 that trains a statistical translation model 1 62, which may be n-gram based or another type of model. As discussed above, the training component 1 60 iterates through source-target pairs of the transcribed queries 1 56 and listings 1 58. In the case of an n-gram based model, given a (source, target) pair, an initial monotonic alignment is obtained between the source form and target form by computing an edit distance. Given the alignment, the training component 1 60 discovers word-level pairs and builds an n-gram translation model 1 62 based on the word- level pairs. The alignment and n-gram model parameters of the translation model 1 62 may be iteratively refined to improve the translation model 1 62. Furthermore, training may implement a backoff strategy which assumes that a word can be translated to itself, as is possible with intra-language translation. In other words, the aligned units WORD-WORD, where WORD can be a word or a phrase, will have a positive probability. [0031 ] A translation module 1 64 uses the translation model 1 62 to test decoded candidates (potential translations). Given the trained translation model 1 62 and a source form, a best-first search algorithm may be used to obtain the top n-best target forms (the n decoded target forms with the highest probability according to the translation model 1 62). The weight of each target form is determined by p(target| source) produced by the translation model. Unlikely word-level pairs may be pruned to speed up translation. [0032] Given the translation model 1 62 and the translation module 1 64, subsequent searches may be improved as follows. Given a user's query q and a listing d found by the search engine 1 52, translated query forms x of the listing d are considered when measuring the listing d's relevancy to a user's query. Letting s(_,_) be a function or measure of relevancy (or similarity), the measurement of relevancy may be s(q, d) = sum_x { p(x |d) s(q, x) }. Alternatively, relevancy may be measured directly from the translation probability, in which case s(q, d) = p(q,d). In one embodiment, potential translations can be filtered out if their similarity measure is below a specified threshold.
[0033] Furthermore, not only may searching be improved as described above, but a language model 1 68 can also be built or augmented using intra-language translation. Language models are used in many natural language processing applications such as ASR, machine translation, and parsing. The intra-language translation provided by the translation model 1 62 and translation module 1 64 may be used in language modeling by translating listings into query forms and using the same-language translated query forms when estimating a language model 1 68. When estimating the language model 1 68, the count of a translated query form may be set to its posterior probabilities multiplied by the count of its original listing.
[0034] In one embodiment, a server- or client-based voice recognizer may be provided with the language model 1 68, which will allow the voice recognizer to perform more accurate and comprehensive speech recognition with respect to utterances directed to the listings 1 54 or to listings. The translation model 1 62 may also be used at a server or at a mobile client to translate a string inputted at the mobile device (whether by ASR or otherwise) to a display form.
CONCLUSION
[0035] Embodiments and features discussed above can be realized in the form of information stored in volatile or non-volatile computer or device readable media. This is deemed to include at least media such as optical storage (e.g., CD-ROM), magnetic media, flash ROM, or any current or future means of storing digital information. The stored information can be in the form of machine executable instructions (e.g., compiled executable binary code), source code, bytecode, or any other information that can be used to enable or configure computing devices to perform the various embodiments discussed above. This is also deemed to include at least volatile memory such as RAM and/or virtual memory storing information such as CPU instructions during execution of a program carrying out an embodiment, as well as non-volatile media storing information that allows a program or executable to be loaded and executed. The embodiments and featured can be performed on any type of computing device, including portable devices, workstations, servers, mobile wireless devices, and so on. The modules, components, processes, and search engine 1 52 discussed above may by realized on one computing device or multiple cooperating computing devices.

Claims

1 . A computer implemented method for intra-language machine translation of phrases in a human language, the method comprising: receiving training data, the training data comprising pairings of source phrases and target phrases (1 20, 1 56, 1 58); using the pairs of training data to train (1 00) an intra-language statistical machine translation model (1 60, 1 62), where the intra-language statistical machine translation model, when given an input phrase of text in the human language, can compute probabilities of semantic equivalence of the input phrase to possible translations of the input phrase in the human language (1 02); and using the statistical machine translation model to translate between queries and listings (1 64), where the queries comprise text strings in the human language submitted to a search engine (1 56), where the listing strings comprise text strings of formal names of real world entities that are to be searched by the search engine to find matches for the query strings (1 58).
2. A method according to claim 1 , wherein the using the intra-language statistical translation model comprises: receiving from the search engine listings that the search engine matched to the user's query (1 58); generating query forms of one of the listings by using the translation model to translate the one of the listings to the query forms (1 64, 1 66); using the translation model to compute similarities of the query forms to the user's query, and determining that the listing does not match the user's query based on the computed similarities.
3. A method according to claim 1 , wherein the using the intra-language statistical translation model comprises: receiving from the search engine a listing that the search engine matched to the user's query (1 58); using the model (1 62) to find a probability that the listing is a translation of the user's query; and determining whether the listing matches the user's query based on the probability.
4. A method according to claim 1 , wherein the intra-language statistical translation model comprises an n-gram based translation model (1 26).
5. A method according to claim 4, wherein the training comprises rearranging either the source and/or target phrase of a training pair so that semantically equivalent words of the source and target phrases are aligned (1 22); and iteratively updating the alignment and parameters of the n-gram based translation model.
6. A method according to claim 1 , further comprising using the intra-language statistical translation model to generate a language model of the human language (1 68), the language model being capable of determining a likelihood of strings in the human language.
7. A method according to claim 6, wherein the using the intra-language statistical translation model to generate the language model comprises including with the language model translations from the intra-language statistical translation model.
8. A method according to claim 6, further comprising performing automatic speech recognition with the language model.
9. One or more computer readable media storing information to enable a computing device to perform a process for translating phrases of a human language to other phrases of the language, the process comprising: accessing training pairs comprising pairs of phrases in the human language (1 20, 1 56, 1 58); training a statistical machine translation model with the training pairs (1 60), the statistical machine translation model being capable of computing probabilities that a target string in the human language is a valid translation of a given source string in the human language; receiving a text phrase in the human language, decoding the text phrase to different candidate translations of the text phrase (1 02), and using the statistical machine translation model to compute probabilities that the candidate translations are translations of the text phrase; and based on the probabilities, storing and/or displaying, by computer, one or more of the candidate translations.
1 0. One or more computer readable media according to claim 9, wherein the training pairs comprise query strings submitted by users paired with listings that a search engine matched with the query strings.
1 1 . One or more computer readable media according to claim 9, wherein the received text phrase comprises a query string inputted by a user, the query string comprising text in the human language, and the process further comprises using the statistical machine translation model to identify a plurality of probable potential translations of the query string, the potential translations comprising text in the human language.
1 2. One or more computer readable media according to claim 9, wherein the received text phrase comprises a name of an organization or business entity obtained from a search engine for searching listings of business/organization names, the name having been obtained from the search engine according to a user-inputted query, and wherein the process further comprises using the statistical machine translation model to determine a probability that the name is a valid translation of the query string and determining relevancy of the listing to the query based on the probability.
1 3. One or more computer readable media according to claim 9, further comprising using the statistical machine translation model to build a statistical language model of the human language, where the statistical language model provides probabilities of phrases in the human language.
14. One or more computer readable media according to claim 9, the process further comprising using the statistical machine translation model to translate text queries recognized by a speech recognizer into display forms.
1 5. A method performed by one or more computers, the method comprising: providing a statistical machine translation model (1 62) configured to compute probabilities of translations of phrases of a same human language; and using the statistical machine translation model to translate between query forms and listing forms of organizations and/or businesses, where the query forms comprise phrases, in the human language, submitted to a search engine (1 52), and where the listing forms comprise formal names, in the human language, of organizations and/or businesses searchable by the search engine.
1 6. A method according to claim 1 5, wherein the using the statistical machine translation model comprises computing similarity between a query form and a listing form.
1 7. A method according to claim 1 5, wherein, given a user query inputted by a user in the human language (1 56), given a corresponding listing in the human language that was found by the search engine (1 52), and given a set of candidate translations of the listing, the candidate translations also in the human language, the using the statistical machine translation model comprises computing probabilities of the candidate translations.
1 8. A method according to claim 1 7, further comprising generating a search result for the given user query based on the computed probabilities.
1 9. A method according to claim 1 5, further comprising generating or modifying a search result of the search engine based on probabilities computed by the statistical machine translation model, the search result corresponding to a user-inputted query form.
20. A method according to claim 1 9, further comprising using the probabilities to rank or eliminate search result.
PCT/US2009/035389 2008-03-28 2009-02-27 Intra-language statistical machine translation WO2009120449A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP09725780.2A EP2269148B1 (en) 2008-03-28 2009-02-27 Intra-language statistical machine translation
CN200980112180.XA CN101981566B (en) 2008-03-28 2009-02-27 Intra-language statistical machine translation

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/058,328 US8615388B2 (en) 2008-03-28 2008-03-28 Intra-language statistical machine translation
US12/058,328 2008-03-28

Publications (1)

Publication Number Publication Date
WO2009120449A1 true WO2009120449A1 (en) 2009-10-01

Family

ID=41114273

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2009/035389 WO2009120449A1 (en) 2008-03-28 2009-02-27 Intra-language statistical machine translation

Country Status (5)

Country Link
US (1) US8615388B2 (en)
EP (1) EP2269148B1 (en)
KR (1) KR101543992B1 (en)
CN (1) CN101981566B (en)
WO (1) WO2009120449A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2388711A3 (en) * 2010-04-26 2012-05-02 Honeywell International, Inc. System and method for translating an English language message into another language
CN103092831A (en) * 2013-01-25 2013-05-08 哈尔滨工业大学 Parameter adjustment method used for counting machine translation
WO2014190220A3 (en) * 2013-05-24 2015-05-14 Microsoft Corporation Language model trained using predicted queries from statistical machine translation
US9830384B2 (en) 2015-10-29 2017-11-28 International Business Machines Corporation Foreign organization name matching
US10949904B2 (en) * 2014-10-04 2021-03-16 Proz.Com Knowledgebase with work products of service providers and processing thereof
US11734600B2 (en) 2018-05-08 2023-08-22 Google Llc Contrastive sequence-to-sequence data selector

Families Citing this family (190)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060116865A1 (en) 1999-09-17 2006-06-01 Www.Uniscape.Com E-services translation utilizing machine translation and translation memory
US7904595B2 (en) 2001-01-18 2011-03-08 Sdl International America Incorporated Globalization management system and method therefor
US7983896B2 (en) 2004-03-05 2011-07-19 SDL Language Technology In-context exact (ICE) matching
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10319252B2 (en) 2005-11-09 2019-06-11 Sdl Inc. Language capability assessment and training apparatus and techniques
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US9002869B2 (en) * 2007-06-22 2015-04-07 Google Inc. Machine translation for query expansion
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
FR2932942B1 (en) * 2008-06-23 2011-03-18 Alcatel Lucent METHOD FOR RECOVERING INFORMATION FROM A TELEPHONE TERMINAL VIA A COMMUNICATION SERVER AND ASSOCIATED COMMUNICATION SERVER
US8301437B2 (en) * 2008-07-24 2012-10-30 Yahoo! Inc. Tokenization platform
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US20100153366A1 (en) * 2008-12-15 2010-06-17 Motorola, Inc. Assigning an indexing weight to a search term
US20100250335A1 (en) * 2009-03-31 2010-09-30 Yahoo! Inc System and method using text features for click prediction of sponsored search advertisements
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
EP2534585A4 (en) * 2010-02-12 2018-01-24 Google LLC Compound splitting
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US10417646B2 (en) 2010-03-09 2019-09-17 Sdl Inc. Predicting the cost associated with translating textual content
US20110295897A1 (en) * 2010-06-01 2011-12-01 Microsoft Corporation Query correction probability based on query-correction pairs
US8612205B2 (en) * 2010-06-14 2013-12-17 Xerox Corporation Word alignment method and system for improved vocabulary coverage in statistical machine translation
KR101682207B1 (en) * 2010-08-23 2016-12-12 에스케이플래닛 주식회사 Apparatus and method for decoding using joint tokenization and translation
US10657540B2 (en) 2011-01-29 2020-05-19 Sdl Netherlands B.V. Systems, methods, and media for web content management
US9547626B2 (en) 2011-01-29 2017-01-17 Sdl Plc Systems, methods, and media for managing ambient adaptability of web applications and web services
US10580015B2 (en) 2011-02-25 2020-03-03 Sdl Netherlands B.V. Systems, methods, and media for executing and optimizing online marketing initiatives
US10140320B2 (en) 2011-02-28 2018-11-27 Sdl Inc. Systems, methods, and media for generating analytical data
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9298287B2 (en) 2011-03-31 2016-03-29 Microsoft Technology Licensing, Llc Combined activation for natural user interface systems
US10642934B2 (en) 2011-03-31 2020-05-05 Microsoft Technology Licensing, Llc Augmented conversational understanding architecture
US9842168B2 (en) 2011-03-31 2017-12-12 Microsoft Technology Licensing, Llc Task driven user intents
US9244984B2 (en) 2011-03-31 2016-01-26 Microsoft Technology Licensing, Llc Location based conversational understanding
US9760566B2 (en) 2011-03-31 2017-09-12 Microsoft Technology Licensing, Llc Augmented conversational understanding agent to identify conversation context between two humans and taking an agent action thereof
US9858343B2 (en) 2011-03-31 2018-01-02 Microsoft Technology Licensing Llc Personalization of queries, conversations, and searches
US9064006B2 (en) 2012-08-23 2015-06-23 Microsoft Technology Licensing, Llc Translating natural language utterances to keyword search queries
US9454962B2 (en) 2011-05-12 2016-09-27 Microsoft Technology Licensing, Llc Sentence simplification for spoken language understanding
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US8838434B1 (en) * 2011-07-29 2014-09-16 Nuance Communications, Inc. Bootstrap call router to other languages using selected N-best translations
US9984054B2 (en) 2011-08-24 2018-05-29 Sdl Inc. Web interface including the review and manipulation of a web document and utilizing permission based control
US9009148B2 (en) 2011-12-19 2015-04-14 Microsoft Technology Licensing, Llc Clickthrough-based latent semantic model
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9773270B2 (en) 2012-05-11 2017-09-26 Fredhopper B.V. Method and system for recommending products based on a ranking cocktail
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US10261994B2 (en) 2012-05-25 2019-04-16 Sdl Inc. Method and system for automatic management of reputation of translators
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US10452740B2 (en) 2012-09-14 2019-10-22 Sdl Netherlands B.V. External content libraries
US11386186B2 (en) 2012-09-14 2022-07-12 Sdl Netherlands B.V. External content library connector systems and methods
US11308528B2 (en) 2012-09-14 2022-04-19 Sdl Netherlands B.V. Blueprinting of multimedia assets
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
US9916306B2 (en) 2012-10-19 2018-03-13 Sdl Inc. Statistical linguistic analysis of source content
US9104733B2 (en) * 2012-11-29 2015-08-11 Microsoft Technology Licensing, Llc Web search ranking
US9183197B2 (en) 2012-12-14 2015-11-10 Microsoft Technology Licensing, Llc Language processing resources for automated mobile language translation
EP2954514B1 (en) 2013-02-07 2021-03-31 Apple Inc. Voice trigger for a digital assistant
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
WO2014200728A1 (en) 2013-06-09 2014-12-18 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9842101B2 (en) * 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9760559B2 (en) * 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
RU2639684C2 (en) * 2014-08-29 2017-12-21 Общество С Ограниченной Ответственностью "Яндекс" Text processing method (versions) and constant machine-readable medium (versions)
US9953646B2 (en) 2014-09-02 2018-04-24 Belleau Technologies Method and system for dynamic speech recognition and tracking of prewritten script
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9817808B2 (en) * 2014-09-29 2017-11-14 International Business Machines Corporation Translation using related term pairs
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
CN111291553B (en) * 2014-10-24 2023-11-21 谷歌有限责任公司 Neural machine translation system with rare word processing
US10452786B2 (en) * 2014-12-29 2019-10-22 Paypal, Inc. Use of statistical flow data for machine translations between different languages
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10545958B2 (en) 2015-05-18 2020-01-28 Microsoft Technology Licensing, Llc Language scaling platform for natural language processing systems
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
CN106484681B (en) 2015-08-25 2019-07-09 阿里巴巴集团控股有限公司 A kind of method, apparatus and electronic equipment generating candidate translation
CN106484682B (en) 2015-08-25 2019-06-25 阿里巴巴集团控股有限公司 Machine translation method, device and electronic equipment based on statistics
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10614167B2 (en) 2015-10-30 2020-04-07 Sdl Plc Translation review workflow systems and methods
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
KR102195627B1 (en) 2015-11-17 2020-12-28 삼성전자주식회사 Apparatus and method for generating translation model, apparatus and method for automatic translation
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179309B1 (en) 2016-06-09 2018-04-23 Apple Inc Intelligent automated assistant in a home environment
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
KR102565274B1 (en) 2016-07-07 2023-08-09 삼성전자주식회사 Automatic interpretation method and apparatus, and machine translation method and apparatus
CN107704456B (en) * 2016-08-09 2023-08-29 松下知识产权经营株式会社 Identification control method and identification control device
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
CN107885716B (en) * 2016-09-29 2020-02-11 腾讯科技(深圳)有限公司 Text recognition method and device
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
CN108304412B (en) * 2017-01-13 2022-09-30 北京搜狗科技发展有限公司 Cross-language search method and device for cross-language search
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK201770428A1 (en) 2017-05-12 2019-02-18 Apple Inc. Low-latency intelligent automated assistant
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US10303715B2 (en) 2017-05-16 2019-05-28 Apple Inc. Intelligent automated assistant for media exploration
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. Far-field extension for digital assistant services
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10635863B2 (en) 2017-10-30 2020-04-28 Sdl Inc. Fragment recall and adaptive automated translation
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10817676B2 (en) 2017-12-27 2020-10-27 Sdl Inc. Intelligent routing services and systems
US10803256B2 (en) 2017-12-29 2020-10-13 Paypal, Inc. Systems and methods for translation management
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10747962B1 (en) 2018-03-12 2020-08-18 Amazon Technologies, Inc. Artificial intelligence system using phrase tables to evaluate and improve neural network based machine translation
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11256867B2 (en) 2018-10-09 2022-02-22 Sdl Inc. Systems and methods of machine learning for digital assets and message creation
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
CN111597800B (en) * 2019-02-19 2023-12-12 百度在线网络技术(北京)有限公司 Method, device, equipment and storage medium for obtaining synonyms
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US20210209315A1 (en) * 2019-03-29 2021-07-08 Google Llc Direct Speech-to-Speech Translation via Machine Learning
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
CN110147558B (en) * 2019-05-28 2023-07-25 北京金山数字娱乐科技有限公司 Method and device for processing translation corpus
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11410667B2 (en) 2019-06-28 2022-08-09 Ford Global Technologies, Llc Hierarchical encoder for speech conversion system
CN110852324A (en) * 2019-08-23 2020-02-28 上海撬动网络科技有限公司 Deep neural network-based container number detection method
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
JP7332486B2 (en) * 2020-01-08 2023-08-23 株式会社東芝 SYMBOL STRING CONVERTER AND SYMBOL STRING CONVERSION METHOD
CN111680527B (en) * 2020-06-09 2023-09-19 语联网(武汉)信息技术有限公司 Man-machine co-interpretation system and method based on dedicated machine turning engine training
US11829720B2 (en) 2020-09-01 2023-11-28 Apple Inc. Analysis and validation of language models

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5991710A (en) * 1997-05-20 1999-11-23 International Business Machines Corporation Statistical translation system with features based on phrases or groups of words
EP1217534A2 (en) * 2000-12-18 2002-06-26 Xerox Corporation Method and apparatus for terminology translation
US20060217963A1 (en) * 2005-03-23 2006-09-28 Fuji Xerox Co., Ltd. Translation memory system
US20070083359A1 (en) * 2003-10-08 2007-04-12 Bender Howard J Relationship analysis system and method for semantic disambiguation of natural language

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11143877A (en) * 1997-10-22 1999-05-28 Internatl Business Mach Corp <Ibm> Compression method, method for compressing entry index data and machine translation system
US20060143007A1 (en) * 2000-07-24 2006-06-29 Koh V E User interaction with voice information services
US6999932B1 (en) * 2000-10-10 2006-02-14 Intel Corporation Language independent voice-based search system
AUPR082400A0 (en) * 2000-10-17 2000-11-09 Telstra R & D Management Pty Ltd An information retrieval system
US6934683B2 (en) * 2001-01-31 2005-08-23 Microsoft Corporation Disambiguation language model
US7027987B1 (en) * 2001-02-07 2006-04-11 Google Inc. Voice interface for a search engine
US6721728B2 (en) * 2001-03-02 2004-04-13 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration System, method and apparatus for discovering phrases in a database
US7295965B2 (en) * 2001-06-29 2007-11-13 Honeywell International Inc. Method and apparatus for determining a measure of similarity between natural language sentences
US6985861B2 (en) * 2001-12-12 2006-01-10 Hewlett-Packard Development Company, L.P. Systems and methods for combining subword recognition and whole word recognition of a spoken input
US7177814B2 (en) * 2002-02-07 2007-02-13 Sap Aktiengesellschaft Dynamic grammar for voice-enabled applications
US7293015B2 (en) * 2002-09-19 2007-11-06 Microsoft Corporation Method and system for detecting user intentions in retrieval of hint sentences
CN100380373C (en) * 2002-10-29 2008-04-09 埃里·阿博 Knowledge system method and apparatus
US7593843B2 (en) * 2004-03-30 2009-09-22 Microsoft Corporation Statistical language model for logical form using transfer mappings
US20060036438A1 (en) * 2004-07-13 2006-02-16 Microsoft Corporation Efficient multimodal method to provide input to a computing device
US20080077570A1 (en) * 2004-10-25 2008-03-27 Infovell, Inc. Full Text Query and Search Systems and Method of Use
US7698124B2 (en) * 2004-11-04 2010-04-13 Microsoft Corporaiton Machine translation system incorporating syntactic dependency treelets into a statistical framework
US7672931B2 (en) * 2005-06-30 2010-03-02 Microsoft Corporation Searching for content using voice search queries
US20070061335A1 (en) * 2005-09-14 2007-03-15 Jorey Ramer Multimodal search query processing
JP4058071B2 (en) * 2005-11-22 2008-03-05 株式会社東芝 Example translation device, example translation method, and example translation program
US7813926B2 (en) * 2006-03-16 2010-10-12 Microsoft Corporation Training system for a speech recognition application
US8943080B2 (en) * 2006-04-07 2015-01-27 University Of Southern California Systems and methods for identifying parallel documents and sentence fragments in multilingual document collections
US7949514B2 (en) * 2007-04-20 2011-05-24 Xerox Corporation Method for building parallel corpora
US20080086298A1 (en) * 2006-10-10 2008-04-10 Anisimovich Konstantin Method and system for translating sentences between langauges
US9002869B2 (en) * 2007-06-22 2015-04-07 Google Inc. Machine translation for query expansion

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5991710A (en) * 1997-05-20 1999-11-23 International Business Machines Corporation Statistical translation system with features based on phrases or groups of words
EP1217534A2 (en) * 2000-12-18 2002-06-26 Xerox Corporation Method and apparatus for terminology translation
US20070083359A1 (en) * 2003-10-08 2007-04-12 Bender Howard J Relationship analysis system and method for semantic disambiguation of natural language
US20060217963A1 (en) * 2005-03-23 2006-09-28 Fuji Xerox Co., Ltd. Translation memory system

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ADAM BERGER ET AL.: "Information Retrieval as Statistical Translation", PROC. SIGIR 1999, 15 August 1999 (1999-08-15), pages 222 - 229, XP002495566
C. QUIRK: "Monolingual Machine Translation for Paraphrase Generation", PROCEEDINGS OF THE CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, 25 July 2004 (2004-07-25), pages 142 - 149, XP002372807
JOSEP M. CREGO; ADRIA DE GISPERT; JOSE B. MARINO., THE TALP NGRAM-BASED SMT SYSTEM FOR IWSLT'05, 24 October 2005 (2005-10-24)
STEFAN RIEZLER ET AL.: "Statistical Machine Translation for Query Expansion in Answer Retrieval", ALC 2007 PROCEEDINGS, 23 June 2007 (2007-06-23)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2388711A3 (en) * 2010-04-26 2012-05-02 Honeywell International, Inc. System and method for translating an English language message into another language
CN103092831A (en) * 2013-01-25 2013-05-08 哈尔滨工业大学 Parameter adjustment method used for counting machine translation
WO2014190220A3 (en) * 2013-05-24 2015-05-14 Microsoft Corporation Language model trained using predicted queries from statistical machine translation
US10949904B2 (en) * 2014-10-04 2021-03-16 Proz.Com Knowledgebase with work products of service providers and processing thereof
US9830384B2 (en) 2015-10-29 2017-11-28 International Business Machines Corporation Foreign organization name matching
US11734600B2 (en) 2018-05-08 2023-08-22 Google Llc Contrastive sequence-to-sequence data selector

Also Published As

Publication number Publication date
EP2269148B1 (en) 2018-09-05
KR20100135757A (en) 2010-12-27
EP2269148A1 (en) 2011-01-05
EP2269148A4 (en) 2011-06-29
CN101981566B (en) 2015-04-15
US8615388B2 (en) 2013-12-24
US20090248422A1 (en) 2009-10-01
CN101981566A (en) 2011-02-23
KR101543992B1 (en) 2015-08-13

Similar Documents

Publication Publication Date Title
US8615388B2 (en) Intra-language statistical machine translation
Ghannay et al. End-to-end named entity and semantic concept extraction from speech
US10346540B2 (en) Self-learning statistical natural language processing for automatic production of virtual personal assistants
US9330661B2 (en) Accuracy improvement of spoken queries transcription using co-occurrence information
CN106997376B (en) Question and answer sentence similarity calculation method based on multi-level features
CN101510222B (en) Multilayer index voice document searching method
US10073673B2 (en) Method and system for robust tagging of named entities in the presence of source or translation errors
US8713037B2 (en) Translation system adapted for query translation via a reranking framework
US8566076B2 (en) System and method for applying bridging models for robust and efficient speech to speech translation
JP5822432B2 (en) Machine translation method and system
WO2018157789A1 (en) Speech recognition method, computer, storage medium, and electronic apparatus
EP3355301B1 (en) Cross-lingual initialization of language models
Chen et al. A discriminative HMM/N-gram-based retrieval approach for Mandarin spoken documents
Kuo et al. A phonetic similarity model for automatic extraction of transliteration pairs
Ture et al. Exploiting representations from statistical machine translation for cross-language information retrieval
Manishina Data-driven natural language generation using statistical machine translation and discriminative learning
Oger et al. On-demand new word learning using world wide web
Xu et al. Cross-lingual language modeling for low-resource speech recognition
Li et al. Language modeling for voice search: a machine translation approach
Justo et al. Integration of complex language models in ASR and LU systems
JPWO2009041220A1 (en) Abbreviation generation apparatus and program, and abbreviation generation method
Wang et al. Language model data filtering via user simulation and dialogue resynthesis.
Chien et al. A spoken‐access approach for chinese text and speech information retrieval
Oger et al. Local Methods for On-Demand Out-of-Vocabulary Word Retrieval.
Feng et al. Query parsing for voice-enabled mobile local search

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200980112180.X

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09725780

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 5588/CHENP/2010

Country of ref document: IN

ENP Entry into the national phase

Ref document number: 20107021217

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2009725780

Country of ref document: EP