WO2003083709A2 - Statistical machine translation - Google Patents

Statistical machine translation Download PDF

Info

Publication number
WO2003083709A2
WO2003083709A2 PCT/US2003/009749 US0309749W WO03083709A2 WO 2003083709 A2 WO2003083709 A2 WO 2003083709A2 US 0309749 W US0309749 W US 0309749W WO 03083709 A2 WO03083709 A2 WO 03083709A2
Authority
WO
WIPO (PCT)
Prior art keywords
chunk
string
syntactic
machine
cause
Prior art date
Application number
PCT/US2003/009749
Other languages
French (fr)
Other versions
WO2003083709A3 (en
Inventor
Philipp Koehn
Kevin Knight
Original Assignee
University Of Southern California
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University Of Southern California filed Critical University Of Southern California
Priority to AU2003222126A priority Critical patent/AU2003222126A1/en
Priority to DE10392450T priority patent/DE10392450T5/en
Priority to JP2003581063A priority patent/JP2005527894A/en
Publication of WO2003083709A2 publication Critical patent/WO2003083709A2/en
Publication of WO2003083709A3 publication Critical patent/WO2003083709A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/42Data-driven translation
    • G06F40/44Statistical methods, e.g. probability models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/42Data-driven translation
    • G06F40/45Example-based machine translation; Alignment

Definitions

  • Machine translation is the automatic translation, for example, using a computer system, from a first language (a "source” language) into another language (a "target” language) .
  • Systems that perform an MT process are said to "decode” the source language into the target language. From an end-user's perspective, the MT process is relatively straight-forward.
  • the MT process receives as input a source sentence (or "string” of words) and after processing the • input sentence, outputs a translated sentence in the target language.
  • One type of MT process is referred to as a statistical MT decoder.
  • Conventional statistical MT decoders may include a language model (LM) and a translation model (TM) .
  • a method includes detecting a syntactic chunk in a first string in a first language, assigning a syntactic label to the detected syntactic chunk in the first string, aligning the detected syntactic chunk in the first string to a syntactic chunk in a second language string, said aligning based on the assigned syntactic label, and translating each word from the first string into a second word corresponding to a possible translation in the second language.
  • One or more of the following features may also be included. Grouping at least two words from the first string based on part of speech identifiers tagged to the at least two words. Defining connections between the detected syntactic chunk in the source string to a chunk in the second string. Determining connections based on a chunk mapping table, the chunk mapping table using predefined connections based on syntactic chunk labels. Defining a connection between the detected chunk from the first string to at least two non-adjacent chunks in the target string. Defining a connection of at least two detected chunks from the source string to a single chunk in the target string.
  • Fig. 1 is a flowchart of a linguistic statistical translation (LST) process.
  • Fig. 2 illustrates an exemplary source and target sentence.
  • Fig. 3 illustrates a sentence-level chunk reordering table corresponding to the sentences of Fig. 2.
  • Fig. 4 illustrates chunk mapping alignment tables corresponding to Figs. 2-3.
  • Fig. 5 illustrates word translations corresponding
  • a statistical MT system as described herein may be modeled as three separate parts: (1) a language model (LM) that assigns a probability P(e) to any target string of words, (2) a translation model (TM) that assigns a probability P(f
  • LM language model
  • TM translation model
  • a conventional MT system may translate source sentences into target sentences by making a series of word-based decisions. The word-based decisions may include a translation decision where each source word is translated to a target word.
  • a mapping (“alignment”) decision may also be performed for each translated word, e.g., mapping multiple source words to a single target word-based on a determined fertility of the translated source word.
  • a re-arrangement (“distortion”) decision may also be performed, e.g., re-arranging the order of words from a source sentence to translated words in the corresponding target sentence.
  • the translation, mapping and distortion decisions are based on weighted probabilities determined during the translation process.
  • Some source sentences pose translational challenges that are not handled well by conventional word-based MT systems. For example, translational challenges include the translation of phrases, restructuring sentences for syntactical reasons, and translations of non-adjacent words into single words or phrases in the target sentence.
  • Fig. 1 depicts a linguistic statistical translation model (LST) process 10 that includes receiving (15) a source sentence to be translated, assigning (20) a "part of speech” (POS) tag(s) for each source word in a source sentence, and detecting (30) syntactic "chunks" included in the source sentence.
  • LST process 10 also includes actions (40) , (50) and (60) that are based, in part, on the assigned POS tag(s) and/or the detected syntactic chunk(s).
  • POS tag(s) and/or syntactic chun (s) in process 10 allows for improved translations of source to target sentences, and, in particular improved string translations of the translational challenges identified previously.
  • a POS tag refers to an identifying symbol that represents a type of word, e.g., a "VVFIN” symbol may be tagged to a finite verb.
  • An exemplary set of POS tags that may be used in process 10 is referred to as the "Penn Treebank Tag set", and described in Mitchell P. Marcus ,- Bea trice Santorlnl , and Mary Ann Marclnklewlcz : "Building a Large Annotated Corpus of English : The Penn Treebank", in Computational Linguistics , Volume 19, Number 2 (June 1993), pp. 313—330 (Special Issue on Using Large Corpora) , which is hereby incorporated by reference.
  • Chunking refers to the grouping of non-recursive verb, noun, prepositional, or other phrases in a sentence. Chunking may include detecting groupings in a source sentence and the output of chunks combinations in a target sentence. The concept of chunking is discussed in Abney, S. (1991) "Parsing by chunks", In Robert Berwick, Steven Abney, and Carol Tenny: Principle-Based Parsing. Kluwer Academic Publishers.
  • LST process 10 includes receiving (15) an input source sentence to be translated, tagging (20) each word in the source sentence with a POS tag, detecting (30) syntactic chunks (e.g., phrases) in each source sentence, sentence-level chunk re-ordering (40), mapping (50) detected source chunks to target chunks in the target sentence, and translating (60) each word from the source to target sentence.
  • the word translations produced by action (60) may be further refined by use of an optional target language model (70) .
  • Fig. 2 shows an exemplary source sentence 100 with each word having an associated POS tag 110-116 generated during action (20) , and detected syntactic chunks 101-105 generated during action (30) .
  • Detected chunks 101-105 also include syntactic labels, e.g., "N, V, N, V and !, respectively.
  • Syntactic labels refer to the syntactic portion of a sentence for the detected chunk, for example, "N” may refer to a base noun phrase, "V” may refer to a verb complex, P” may refer to a base prepositional phrase, "A” may refer to an adjective, "F” may refer to a function word, and " ! " may refer to punctuation.
  • Sentence-level chunk re-ordering defines connections 120-125 between each source chunk 101-106 and a corresponding target chunk 130-134 that will be included in the target sentence 150.
  • the target chunks are re-ordered relative to the source chunks. This re-ordering may be based upon templates that define likely connections between detected syntactic chunk to corresponding syntactic chunks in a target sentence.
  • Connection (s) may single value or multi-valued (e.g., one-to-one, many-to-many, or one-to-many, etc.).
  • Fig. 3 shows a chunk connection table 160 that is representative of the connections 120-125 between source chunks 101-105 and target chunks 130-134, corresponding to those shown in Fig. 2.
  • Fig. 4 shows chunk mapping tables 170, 180, 190 and 200 that are representative of the chunk mappings created by action (50) of process 10 as applied to exemplary sentence 100.
  • Chunk mapping refers to the alignment of each source chunk to a target chunk and may be referenced in terms of the POS tag of words in the source chunk and words in the target chunk. For example, as shown in table 170, source POS tags 110 ("ART") and 111 ("NN”) are aligned to target POS tags 140 ("DT") and 141 (“NNP”) . Chunk mappings may align multiple chunks
  • Non-adjacent chunks from the source sentence 110 may be combined into a single chunk, for example, as shown in table 180, combining chunks 102 and
  • each complex chunk may be "labeled" with an assigned syntactic chunk label. This labeling may allow improved reordering of chunks at the sentence level, since the syntactic label may identify their syntactic role in a sentence.
  • Process 10 then translates (60) the source words from the source language sentence to words for the target language sentence. Word translation may be determined, in part, on the part-of-speech assigned to the corresponding source word (selected by the chunk mapping), e.g., restricting a selection of a word corresponding to the assigned POS tag.
  • Fig. 5 depicts the performance of action (60) from process 10, e.g., depicting word translations corresponding to the example shown in Figs. 1-4.
  • a complex chunk may be translated by exact phrase lookup.
  • the entire source chunk may be translated as the known phrase. For example, as shown in Fig. 2, if the words included in source chunk 103 "der Agrarausschuss" is a known phrase, it may be translated directly into the words in target chunks 130-131 "the sub-committee for agriculture". Exact phrase lookup allows for the translation of idiomatic phrases that are not easily translated used word-based translations.
  • Process 10 may include an optional target language model (70) that is performed to provide additional fluency improvements to the target sentence.
  • LST process 10 may be modeled mathematically, for example, modeled on a set of probability determinations.
  • the mathematical model of process 10 that follows includes a formulation that follows the noisy channel model. In more detail, this means that instead of estimating p(e
  • the translation part is decomposed into sentence level reordering (SLR) , chunk mapping (CM) and word translations (W) , and may be modeled with the following probability equation:
  • p(SLR) may be conditioned only on each target chunk label sequence
  • p(CMi) may be conditioned only on the relevant source and target chunk labels, and the target POS tags;
  • p(Wi j ) may be conditioned only on the relevant target POS tag and word.
  • Each word alignment in a chunk mapping is factored in with a word translation probability.
  • Unaligned source words are factored in with the probability p (f k
  • Unaligned target words are factored in with the probability p (NULL
  • Parameters for the word alignments may be determined using a so-called parallel corpus method in which text in a source language string (a first corpus) is aligned to translated text in a target language string (the second corpus) . These alignments establish correspondences between source words in the source string and the target string. Both sides of the parallel corpus may also be POS tagged and chunked.
  • Chunk mappings may be determined using a parallel corpus method, for example, if a source chunk and a target chunk contain a source word and a target word that are aligned to each other, the two chunks are connected. Chunks that contain no aligned words may be attached to other chunks based on a set of rules, for example, adverbs are attached to a following verb chunk if they are unaligned, or commas are attached to a following function word if they are unaligned, and so forth.
  • a transitive closure may then be performed on any chunk alignments, for example, using the following rule set: If chunk fi is aligned with e x , f j is aligned with e x , and chunk fi is aligned with e y , then chunk f j is considered aligned with e y , even if they do not contain any words aligned to each other.
  • the transitive closure ensures a one-to-one mapping between complex chunks in a source sentence and a target sentence.
  • the translation part of the model (e.g., "decoding") may be performed in two steps: First a sentence level template (SLT) for each sentence level chunk reordering is generated. Second, a target translation is constructed a word at a time from left to right. This is repeated for the top n SLT for each given source chunk sequence. Ultimately, the translation with the overall best score is selected as a system output.
  • SLT sentence level template
  • the construction of a target sentence for a given sentence level template (SLT) may be implemented by a Viterbi search using dynamic programming. In this case, chunk mapping templates are selected as needed. Then word slots are filled with use of a word-by-word translation table and a language model.
  • the construction of the target string may include the insertion of a NULL word.
  • the following information is maintained: [0045] - last two words created (needed by the language model) ;
  • Stacked chunk mapping templates refers to information needed when a disconnected complex chunk is filled in the target translation: for example, if the SLT calls for the creation of a "V+P" chunk, with additional material between "V” and "P". In this case, the information about the chunk mapping template that was selected has to be carried through between the "V” and "P", until it is completely filled.
  • the complexity of the hypothesis space at any given position in a target sentence may be represented as 0(V 2 C 1+s ), with V being the vocabulary size ' , C the number of applicable chunk mapping templates, and s the number of stacked chunk mapping templates.
  • the model may be simplified by restricting translations to contiguous complex chunks in the target language, which eliminates the need for stacked chunk mapping templates. This simplifies the complexity equation to 0(V 2 C)at any given position. This also assures that, in respect to sentence length, decoding has linear complexity.
  • Fig. 6 shows an embodiment of an LST process 100 that is modeled upon the equations and formulations discussed previously.
  • LST process 100 includes a loop (135, 140, 150, 160 and 170) that is repeated n times for n different sentence level templates .
  • chunk mapping errors may be caused by source words that translate into multiple target words. This type of error may be avoided or reduced by adding fertility features, or further pre-processing of compound nouns.
  • word translation may be performed by use of a probabilistic word translation method, e.g., a "T-Table" translation method.
  • SLT sentence level templates
  • other estimations may be used, for example, clause level templates, or using a method that decomposes the sentence level chunk translation step into a number of chunk segmentation and translation decisions.

Abstract

A method includes detecting a syntactic chunk in a source string in a first language, assigning a syntactic label to the detected syntactic chunk in the source string, mapping the detected syntactic chunk in the source string to a syntactic chunk in a target string in a second language, said mapping based on the assigned syntactic label, and translating the source string into a possible translation in the second language.

Description

STATISTICAL MACHINE TRANSLATION
CROSS-REFERENCE TO RELATED APPLICATIONS [0001] This application claims priority to U.S. Provisional Application Serial No. 60/368,851, filed on March 28, 2002, which is hereby incorporated by reference.
ORIGIN OF INVENTION [0002] The research and development described in this application were supported by DARPA-ITO under grant number N66001-00-1-8914. The U.S. Government may have certain rights in the claimed inventions.,
BACKGROUND [0003] Machine translation (MT) is the automatic translation, for example, using a computer system, from a first language (a "source" language) into another language (a "target" language) . Systems that perform an MT process are said to "decode" the source language into the target language. From an end-user's perspective, the MT process is relatively straight-forward. The MT process receives as input a source sentence (or "string" of words) and after processing the • input sentence, outputs a translated sentence in the target language. [0004] One type of MT process is referred to as a statistical MT decoder. Conventional statistical MT decoders may include a language model (LM) and a translation model (TM) .
SUMMARY
[0005] According to an aspect of this invention, a method includes detecting a syntactic chunk in a first string in a first language, assigning a syntactic label to the detected syntactic chunk in the first string, aligning the detected syntactic chunk in the first string to a syntactic chunk in a second language string, said aligning based on the assigned syntactic label, and translating each word from the first string into a second word corresponding to a possible translation in the second language.
[0006] One or more of the following features may also be included. Grouping at least two words from the first string based on part of speech identifiers tagged to the at least two words. Defining connections between the detected syntactic chunk in the source string to a chunk in the second string. Determining connections based on a chunk mapping table, the chunk mapping table using predefined connections based on syntactic chunk labels. Defining a connection between the detected chunk from the first string to at least two non-adjacent chunks in the target string. Defining a connection of at least two detected chunks from the source string to a single chunk in the target string.
BRIEF DESCRIPTION OF THE DRAWINGS [0007] Fig. 1 is a flowchart of a linguistic statistical translation (LST) process.
[0008] Fig. 2 illustrates an exemplary source and target sentence.
[0009] Fig. 3 illustrates a sentence-level chunk reordering table corresponding to the sentences of Fig. 2. [0010] Fig. 4 illustrates chunk mapping alignment tables corresponding to Figs. 2-3.
[0011] Fig. 5 illustrates word translations corresponding
to Figs. 1-4. [0012] Fig. 6 shows a second embodiment of a LST process. DETAILED DESCRIPTION [0013] A statistical MT system as described herein may be modeled as three separate parts: (1) a language model (LM) that assigns a probability P(e) to any target string of words, (2) a translation model (TM) that assigns a probability P(f|e) to any pair of target and source strings, and (3) a decoder that determines translations based on the assigned probabilities of the LM and TM. [0014] A conventional MT system may translate source sentences into target sentences by making a series of word-based decisions. The word-based decisions may include a translation decision where each source word is translated to a target word. A mapping ("alignment") decision may also be performed for each translated word, e.g., mapping multiple source words to a single target word-based on a determined fertility of the translated source word. A re-arrangement ("distortion") decision may also be performed, e.g., re-arranging the order of words from a source sentence to translated words in the corresponding target sentence. The translation, mapping and distortion decisions are based on weighted probabilities determined during the translation process. [0015] Some source sentences pose translational challenges that are not handled well by conventional word-based MT systems. For example, translational challenges include the translation of phrases, restructuring sentences for syntactical reasons, and translations of non-adjacent words into single words or phrases in the target sentence.
[0016] Fig. 1 depicts a linguistic statistical translation model (LST) process 10 that includes receiving (15) a source sentence to be translated, assigning (20) a "part of speech" (POS) tag(s) for each source word in a source sentence, and detecting (30) syntactic "chunks" included in the source sentence. LST process 10 also includes actions (40) , (50) and (60) that are based, in part, on the assigned POS tag(s) and/or the detected syntactic chunk(s). The use of POS tag(s) and/or syntactic chun (s) in process 10 allows for improved translations of source to target sentences, and, in particular improved string translations of the translational challenges identified previously. [0017] A POS tag refers to an identifying symbol that represents a type of word, e.g., a "VVFIN" symbol may be tagged to a finite verb. An exemplary set of POS tags that may be used in process 10 is referred to as the "Penn Treebank Tag set", and described in Mitchell P. Marcus ,- Bea trice Santorlnl , and Mary Ann Marclnklewlcz : "Building a Large Annotated Corpus of English : The Penn Treebank", in Computational Linguistics , Volume 19, Number 2 (June 1993), pp. 313—330 (Special Issue on Using Large Corpora) , which is hereby incorporated by reference.
[0018] Chunking refers to the grouping of non-recursive verb, noun, prepositional, or other phrases in a sentence. Chunking may include detecting groupings in a source sentence and the output of chunks combinations in a target sentence. The concept of chunking is discussed in Abney, S. (1991) "Parsing by chunks", In Robert Berwick, Steven Abney, and Carol Tenny: Principle-Based Parsing. Kluwer Academic Publishers.
[0019] Still referring to Fig. 1, LST process 10 includes receiving (15) an input source sentence to be translated, tagging (20) each word in the source sentence with a POS tag, detecting (30) syntactic chunks (e.g., phrases) in each source sentence, sentence-level chunk re-ordering (40), mapping (50) detected source chunks to target chunks in the target sentence, and translating (60) each word from the source to target sentence. The word translations produced by action (60) may be further refined by use of an optional target language model (70) . [0020] Fig. 2 shows an exemplary source sentence 100 with each word having an associated POS tag 110-116 generated during action (20) , and detected syntactic chunks 101-105 generated during action (30) . Detected chunks 101-105 also include syntactic labels, e.g., "N, V, N, V and !", respectively. Syntactic labels refer to the syntactic portion of a sentence for the detected chunk, for example, "N" may refer to a base noun phrase, "V" may refer to a verb complex, P" may refer to a base prepositional phrase, "A" may refer to an adjective, "F" may refer to a function word, and " ! " may refer to punctuation.
[0021] Sentence-level chunk re-ordering (40) defines connections 120-125 between each source chunk 101-106 and a corresponding target chunk 130-134 that will be included in the target sentence 150. In many cases, the target chunks are re-ordered relative to the source chunks. This re-ordering may be based upon templates that define likely connections between detected syntactic chunk to corresponding syntactic chunks in a target sentence. Connection (s) may single value or multi-valued (e.g., one-to-one, many-to-many, or one-to-many, etc.). Fig. 3 shows a chunk connection table 160 that is representative of the connections 120-125 between source chunks 101-105 and target chunks 130-134, corresponding to those shown in Fig. 2.
[0022] Fig. 4 shows chunk mapping tables 170, 180, 190 and 200 that are representative of the chunk mappings created by action (50) of process 10 as applied to exemplary sentence 100. Chunk mapping refers to the alignment of each source chunk to a target chunk and may be referenced in terms of the POS tag of words in the source chunk and words in the target chunk. For example, as shown in table 170, source POS tags 110 ("ART") and 111 ("NN") are aligned to target POS tags 140 ("DT") and 141 ("NNP") . Chunk mappings may align multiple chunks
("complex chunks") to single chunks or other complex chunks. For example, as shown in table 190, source chunk
103 is aligned to a complex chunk including target chunk 130 and 131. Non-adjacent chunks from the source sentence 110 may be combined into a single chunk, for example, as shown in table 180, combining chunks 102 and
104 into target chunk 132.
[0023] As described previously, each complex chunk may be "labeled" with an assigned syntactic chunk label. This labeling may allow improved reordering of chunks at the sentence level, since the syntactic label may identify their syntactic role in a sentence. [0024] Process 10 then translates (60) the source words from the source language sentence to words for the target language sentence. Word translation may be determined, in part, on the part-of-speech assigned to the corresponding source word (selected by the chunk mapping), e.g., restricting a selection of a word corresponding to the assigned POS tag. Fig. 5 depicts the performance of action (60) from process 10, e.g., depicting word translations corresponding to the example shown in Figs. 1-4.
[0025] In an embodiment, instead of generating target language words by individual word translations, a complex chunk may be translated by exact phrase lookup. In more detail, if an entire source chunk is determined to be a known phrase, the entire source chunk may be translated as the known phrase. For example, as shown in Fig. 2, if the words included in source chunk 103 "der Agrarausschuss" is a known phrase, it may be translated directly into the words in target chunks 130-131 "the sub-committee for agriculture". Exact phrase lookup allows for the translation of idiomatic phrases that are not easily translated used word-based translations. [0026] Process 10 may include an optional target language model (70) that is performed to provide additional fluency improvements to the target sentence.
[0027] Mathematical Formulation of Process 10 [0028] The operations of LST process 10 may be modeled mathematically, for example, modeled on a set of probability determinations. The mathematical model of process 10 that follows includes a formulation that follows the noisy channel model. In more detail, this means that instead of estimating p(e|f) directly (e.g., the best translation e for an input string f) , Bayes rule is applied to maximize p (f"| e) x p(e). Therefore, this splits the model into two parts: a translation part p(f|e) and a language model p(e). For the language part, a trigram language model may be used.
[0029] The translation part is decomposed into sentence level reordering (SLR) , chunk mapping (CM) and word translations (W) , and may be modeled with the following probability equation:
p(f|e) = p(SLR|e) x πxp (CM± I e, SLR) x IIjP (W±j | CM±, SLR, e) [0030] Since POS tagging and chunking is deterministic, e represents not only words of the target string, but also their POS and groupings into chunks. The sentence level chunk reordering (SLR) and word reordering within chunks (CM) may be performed using templates, for example, using templates representative of the information from the tables shown in Figs. 3 and 4. Word translation (W) may be accomplished using a word-by-word translation table. [0031] Direct application of the three probability equations above may be problematic due to sparse data. Therefore the three conditional probability distributions may be simplified, as follows:
[0032] p(SLR) may be conditioned only on each target chunk label sequence;
[0033] p(CMi) may be conditioned only on the relevant source and target chunk labels, and the target POS tags; [0034] p(Wij) may be conditioned only on the relevant target POS tag and word.
[0035] Each word alignment in a chunk mapping is factored in with a word translation probability. Unaligned source words are factored in with the probability p (fk| ZFERT, fpos ) • Unaligned target words are factored in with the probability p (NULL | ek, fpoSk) •
[0036] Instead of decomposing the chunk mapping into word translations, a direct phrase lookup may be performed, which is modeled by the following equation:
[0037] p(Wil,...,Win|CMi,SLR,e)
[0038] Parameters for the word alignments may be determined using a so-called parallel corpus method in which text in a source language string (a first corpus) is aligned to translated text in a target language string (the second corpus) . These alignments establish correspondences between source words in the source string and the target string. Both sides of the parallel corpus may also be POS tagged and chunked.
[0039] Chunk mappings may be determined using a parallel corpus method, for example, if a source chunk and a target chunk contain a source word and a target word that are aligned to each other, the two chunks are connected. Chunks that contain no aligned words may be attached to other chunks based on a set of rules, for example, adverbs are attached to a following verb chunk if they are unaligned, or commas are attached to a following function word if they are unaligned, and so forth. [0040] A transitive closure may then be performed on any chunk alignments, for example, using the following rule set: If chunk fi is aligned with ex, fj is aligned with ex, and chunk fi is aligned with ey, then chunk fj is considered aligned with ey, even if they do not contain any words aligned to each other. The transitive closure ensures a one-to-one mapping between complex chunks in a source sentence and a target sentence. [0041] Aligning a parallel corpus based on the above formulations allows statistics to be collected on word translations (including p (fk| ZFERT, fposk) and p (NULL I ek, fposk) ) i complex chunk mappings, and sentence level reordering. Conditional probability distributions may then be collected by maximum likelihood estimation. Since the data for exact phrase lookup is highly noisy, the probabilities may be smoothed.
[0042] In an embodiment, the translation part of the model (e.g., "decoding") may be performed in two steps: First a sentence level template (SLT) for each sentence level chunk reordering is generated. Second, a target translation is constructed a word at a time from left to right. This is repeated for the top n SLT for each given source chunk sequence. Ultimately, the translation with the overall best score is selected as a system output. [0043] The construction of a target sentence for a given sentence level template (SLT) may be implemented by a Viterbi search using dynamic programming. In this case, chunk mapping templates are selected as needed. Then word slots are filled with use of a word-by-word translation table and a language model. At the end of each complex chunk, information about which chunk mapping template was used may be discarded. In some implementations the construction of the target string may include the insertion of a NULL word. [0044] However, for each partial translation (or hypothesis) , the following information is maintained: [0045] - last two words created (needed by the language model) ;
[0046] - current chunk mapping template, if not complete; [0047] - current score (''score'' refers to the product of the combined probabilities of partial translation decision, chunk mapping decisions, etc.) [0048] - back pointer to best path; [0049] - position of last chunk;
[0050] - position of last word created within chunk; [0051] - "stacked chunk mapping templates" [0052] Stacked chunk mapping templates refers to information needed when a disconnected complex chunk is filled in the target translation: for example, if the SLT calls for the creation of a "V+P" chunk, with additional material between "V" and "P". In this case, the information about the chunk mapping template that was selected has to be carried through between the "V" and "P", until it is completely filled.
[0053] The complexity of the hypothesis space at any given position in a target sentence may be represented as 0(V2C1+s), with V being the vocabulary size', C the number of applicable chunk mapping templates, and s the number of stacked chunk mapping templates. [0054] The model may be simplified by restricting translations to contiguous complex chunks in the target language, which eliminates the need for stacked chunk mapping templates. This simplifies the complexity equation to 0(V2C)at any given position. This also assures that, in respect to sentence length, decoding has linear complexity.
[0055] Fig. 6 shows an embodiment of an LST process 100 that is modeled upon the equations and formulations discussed previously. In this example, LST process 100 includes a loop (135, 140, 150, 160 and 170) that is repeated n times for n different sentence level templates .
[0056] A number of embodiments have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. For example, chunk mapping errors may be caused by source words that translate into multiple target words. This type of error may be avoided or reduced by adding fertility features, or further pre-processing of compound nouns. As another example, word translation may be performed by use of a probabilistic word translation method, e.g., a "T-Table" translation method. As another example, there may be not sufficient statistics to reliably estimate sentence level templates (SLT) . Therefore, other estimations may be used, for example, clause level templates, or using a method that decomposes the sentence level chunk translation step into a number of chunk segmentation and translation decisions.
[0057] Accordingly, other embodiments are within the scope of the following claims.

Claims

1. A method comprising: detecting a syntactic chunk in a source string in a first language; assigning a syntactic label to the detected syntactic chunk in the source string; mapping the detected syntactic chunk in the source string to a syntactic chunk in a target string in a second language, said mapping based on the assigned syntactic label; and translating the source string into a possible translation in the second language.
2. The method of claim 1, wherein assigning the syntactic label comprises assigning based on a part of speech identifier tagged to a word in the source string.
3. The method of claim 1, further comprises defining connections between the detected syntactic chunk in the source string to a chunk in the target string.
4. The method of claim 3, wherein defining connections comprises determining connections based on a chunk mapping table, the chunk mapping table using predefined connections based on syntactic chunk labels.
5. The method of claim 3, wherein defining connections comprises defining a connection between the detected chunk from the source string to at least two non-adjacent chunks in the target string.
6. The method of claim 3, wherein defining connections comprises defining a connection of at least two detected chunks from the source string to a single chunk in the target string.
7. The method of claim 1, wherein translating comprises including at least two words in the target string that corresponds to a single word in the source string.
8. The method of claim 1, wherein translating comprises translating phrases.
9. The method of claim 1, further comprises: applying a language model to the source string, the language model based upon the language of the target string
10. The method of claim 1, further comprises: determining a probability of said mapping.
11. The method of claim 1, wherein translating comprises inserting at least one NULL word in the target string.
12. An article comprising a machine readable medium including machine-executable instructions, the instructions operative to cause a machine to: detect a syntactic chunk in a source string in a first language; assign a syntactic label to the detected syntactic chunk in the source string; align the detected syntactic chunk in the source string to a syntactic chunk in a second language string, said mapping based on the assigned syntactic label; and translate each word from the source string into a second word corresponding to a possible translation in the second language.
13. The article of claim 12, wherein instructions operative to cause a machine to assign the syntactic label comprises instructions operative to cause a machine to assign the syntactic label based on a part of speech identifier tagged to a word in the source string.
14. The article of claim 12, further comprises instructions operative to cause a machine to define connections between the detected syntactic chunk in the source string to a chunk in the target string.
15. The article of claim 14, wherein instructions operative to cause a machine to define connections comprises instructions operative to cause a machine to determine connections based on a chunk mapping table, the chunk mapping table using pre-defined connections based on syntactic chunk labels .
16. The article of claim 14, wherein instructions operative to cause a machine to define connections comprises instructions operative to cause a machine to define a connection between the detected chunk from the source string to at least two non-adjacent chunks in the target string.
17. The article of claim 3, wherein instructions operative to cause a machine to define connections comprises instructions operative to cause a machine to define a connection of at least two detected chunks from the source string to a single chunk in the target string.
18. The article of claim 12, wherein instructions operative to cause a machine to translate comprises instructions operative to cause a machine to include at least two words in the target string that corresponds to a single word in the source string.
19. The article of claim 12, wherein instructions operative to cause a machine to translate comprises instructions operative to cause a machine to translate phrases .
20. The article of claim 12, further comprises instructions operative to cause a machine to: apply a language model to the source string, the language model based upon the language of the target string
21. The article of claim 12, further comprises instructions operative to cause a machine to: determine a probability of said mapping.
22. The article of claim 12, wherein instructions operative to cause a machine to translate comprises instructions operative to cause a machine to insert at least one NULL word in the target string.
PCT/US2003/009749 2002-03-28 2003-03-28 Statistical machine translation WO2003083709A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
AU2003222126A AU2003222126A1 (en) 2002-03-28 2003-03-28 Statistical machine translation
DE10392450T DE10392450T5 (en) 2002-03-28 2003-03-28 Statistical machine translation
JP2003581063A JP2005527894A (en) 2002-03-28 2003-03-28 Statistical machine translation

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US36885102P 2002-03-28 2002-03-28
US60/368,851 2002-03-28

Publications (2)

Publication Number Publication Date
WO2003083709A2 true WO2003083709A2 (en) 2003-10-09
WO2003083709A3 WO2003083709A3 (en) 2004-09-10

Family

ID=28675546

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2003/009749 WO2003083709A2 (en) 2002-03-28 2003-03-28 Statistical machine translation

Country Status (6)

Country Link
US (1) US7624005B2 (en)
JP (1) JP2005527894A (en)
CN (1) CN1647068A (en)
AU (1) AU2003222126A1 (en)
DE (1) DE10392450T5 (en)
WO (1) WO2003083709A2 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8886517B2 (en) 2005-06-17 2014-11-11 Language Weaver, Inc. Trust scoring for language translation systems
US8886518B1 (en) 2006-08-07 2014-11-11 Language Weaver, Inc. System and method for capitalizing machine translated text
US8942973B2 (en) 2012-03-09 2015-01-27 Language Weaver, Inc. Content page URL translation
US8990064B2 (en) 2009-07-28 2015-03-24 Language Weaver, Inc. Translating documents based on content
US10261994B2 (en) 2012-05-25 2019-04-16 Sdl Inc. Method and system for automatic management of reputation of translators
US10319252B2 (en) 2005-11-09 2019-06-11 Sdl Inc. Language capability assessment and training apparatus and techniques
US10417646B2 (en) 2010-03-09 2019-09-17 Sdl Inc. Predicting the cost associated with translating textual content
US11003838B2 (en) 2011-04-18 2021-05-11 Sdl Inc. Systems and methods for monitoring post translation editing

Families Citing this family (83)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060116865A1 (en) 1999-09-17 2006-06-01 Www.Uniscape.Com E-services translation utilizing machine translation and translation memory
US7904595B2 (en) 2001-01-18 2011-03-08 Sdl International America Incorporated Globalization management system and method therefor
US8214196B2 (en) 2001-07-03 2012-07-03 University Of Southern California Syntax-based statistical translation model
WO2004001623A2 (en) 2002-03-26 2003-12-31 University Of Southern California Constructing a translation lexicon from comparable, non-parallel corpora
US8548794B2 (en) * 2003-07-02 2013-10-01 University Of Southern California Statistical noun phrase translation
US7711545B2 (en) * 2003-07-02 2010-05-04 Language Weaver, Inc. Empirical methods for splitting compound words with application to machine translation
JP2005100335A (en) * 2003-09-01 2005-04-14 Advanced Telecommunication Research Institute International Machine translation apparatus, machine translation computer program, and computer
JP3919771B2 (en) * 2003-09-09 2007-05-30 株式会社国際電気通信基礎技術研究所 Machine translation system, control device thereof, and computer program
US7587307B2 (en) * 2003-12-18 2009-09-08 Xerox Corporation Method and apparatus for evaluating machine translation quality
US7983896B2 (en) 2004-03-05 2011-07-19 SDL Language Technology In-context exact (ICE) matching
US8296127B2 (en) 2004-03-23 2012-10-23 University Of Southern California Discovery of parallel text portions in comparable collections of corpora and training using comparable texts
US8666725B2 (en) 2004-04-16 2014-03-04 University Of Southern California Selection and use of nonstatistical translation components in a statistical machine translation framework
US20070016401A1 (en) * 2004-08-12 2007-01-18 Farzad Ehsani Speech-to-speech translation system with user-modifiable paraphrasing grammars
JP5452868B2 (en) 2004-10-12 2014-03-26 ユニヴァーシティー オブ サザン カリフォルニア Training for text-to-text applications that use string-to-tree conversion for training and decoding
US8676563B2 (en) 2009-10-01 2014-03-18 Language Weaver, Inc. Providing human-generated and machine-generated trusted translations
WO2006133571A1 (en) * 2005-06-17 2006-12-21 National Research Council Of Canada Means and method for adapted language translation
US20070010989A1 (en) * 2005-07-07 2007-01-11 International Business Machines Corporation Decoding procedure for statistical machine translation
US7536295B2 (en) * 2005-12-22 2009-05-19 Xerox Corporation Machine translation using non-contiguous fragments of text
US8943080B2 (en) 2006-04-07 2015-01-27 University Of Southern California Systems and methods for identifying parallel documents and sentence fragments in multilingual document collections
US7827028B2 (en) * 2006-04-07 2010-11-02 Basis Technology Corporation Method and system of machine translation
US8209162B2 (en) * 2006-05-01 2012-06-26 Microsoft Corporation Machine translation split between front end and back end processors
US9020804B2 (en) 2006-05-10 2015-04-28 Xerox Corporation Method for aligning sentences at the word level enforcing selective contiguity constraints
US7542893B2 (en) * 2006-05-10 2009-06-02 Xerox Corporation Machine translation using elastic chunks
JP5082374B2 (en) * 2006-10-19 2012-11-28 富士通株式会社 Phrase alignment program, translation program, phrase alignment device, and phrase alignment method
US8433556B2 (en) 2006-11-02 2013-04-30 University Of Southern California Semi-supervised training for statistical word alignment
US9122674B1 (en) 2006-12-15 2015-09-01 Language Weaver, Inc. Use of annotations in statistical machine translation
US8468149B1 (en) 2007-01-26 2013-06-18 Language Weaver, Inc. Multi-lingual online community
US8615389B1 (en) 2007-03-16 2013-12-24 Language Weaver, Inc. Generation and exploitation of an approximate language model
US8831928B2 (en) 2007-04-04 2014-09-09 Language Weaver, Inc. Customizable machine translation service
US9779079B2 (en) * 2007-06-01 2017-10-03 Xerox Corporation Authoring system
US8825466B1 (en) 2007-06-08 2014-09-02 Language Weaver, Inc. Modification of annotated bilingual segment pairs in syntax-based machine translation
US8452585B2 (en) * 2007-06-21 2013-05-28 Microsoft Corporation Discriminative syntactic word order model for machine translation
US8046211B2 (en) 2007-10-23 2011-10-25 Microsoft Corporation Technologies for statistical machine translation based on generated reordering knowledge
US8275607B2 (en) * 2007-12-12 2012-09-25 Microsoft Corporation Semi-supervised part-of-speech tagging
US8150677B2 (en) * 2008-06-26 2012-04-03 Microsoft Corporation Machine translation using language order templates
US9176952B2 (en) * 2008-09-25 2015-11-03 Microsoft Technology Licensing, Llc Computerized statistical machine translation with phrasal decoder
US8407042B2 (en) * 2008-12-09 2013-03-26 Xerox Corporation Cross language tool for question answering
US8280718B2 (en) * 2009-03-16 2012-10-02 Xerox Corporation Method to preserve the place of parentheses and tags in statistical machine translation systems
US8326599B2 (en) * 2009-04-21 2012-12-04 Xerox Corporation Bi-phrase filtering for statistical machine translation
US8380486B2 (en) 2009-10-01 2013-02-19 Language Weaver, Inc. Providing machine-generated translations and corresponding trust levels
US8548796B2 (en) * 2010-01-20 2013-10-01 Xerox Corporation Statistical machine translation system and method for translation of text into languages which produce closed compound words
US9552355B2 (en) 2010-05-20 2017-01-24 Xerox Corporation Dynamic bi-phrases for statistical machine translation
US8612205B2 (en) 2010-06-14 2013-12-17 Xerox Corporation Word alignment method and system for improved vocabulary coverage in statistical machine translation
US20120035905A1 (en) 2010-08-09 2012-02-09 Xerox Corporation System and method for handling multiple languages in text
KR101745349B1 (en) * 2010-09-02 2017-06-09 에스케이플래닛 주식회사 Apparatus and method for fiding general idiomatic expression using phrase alignment of parallel corpus
US8775155B2 (en) 2010-10-25 2014-07-08 Xerox Corporation Machine translation using overlapping biphrase alignments and sampling
TWI434187B (en) * 2010-11-03 2014-04-11 Inst Information Industry Text conversion method and system
CN102486770B (en) * 2010-12-02 2014-09-17 财团法人资讯工业策进会 Character conversion method and system
US10657540B2 (en) 2011-01-29 2020-05-19 Sdl Netherlands B.V. Systems, methods, and media for web content management
US9547626B2 (en) 2011-01-29 2017-01-17 Sdl Plc Systems, methods, and media for managing ambient adaptability of web applications and web services
US10580015B2 (en) 2011-02-25 2020-03-03 Sdl Netherlands B.V. Systems, methods, and media for executing and optimizing online marketing initiatives
US10140320B2 (en) 2011-02-28 2018-11-27 Sdl Inc. Systems, methods, and media for generating analytical data
US8798984B2 (en) 2011-04-27 2014-08-05 Xerox Corporation Method and system for confidence-weighted learning of factored discriminative language models
US20120303352A1 (en) * 2011-05-24 2012-11-29 The Boeing Company Method and apparatus for assessing a translation
WO2012170817A1 (en) * 2011-06-10 2012-12-13 Google Inc. Augmenting statistical machine translation with linguistic knowledge
US8694303B2 (en) 2011-06-15 2014-04-08 Language Weaver, Inc. Systems and methods for tuning parameters in statistical machine translation
US8713037B2 (en) * 2011-06-30 2014-04-29 Xerox Corporation Translation system adapted for query translation via a reranking framework
US8781810B2 (en) 2011-07-25 2014-07-15 Xerox Corporation System and method for productive generation of compound words in statistical machine translation
US9984054B2 (en) 2011-08-24 2018-05-29 Sdl Inc. Web interface including the review and manipulation of a web document and utilizing permission based control
US8886515B2 (en) 2011-10-19 2014-11-11 Language Weaver, Inc. Systems and methods for enhancing machine translation post edit review processes
US9773270B2 (en) 2012-05-11 2017-09-26 Fredhopper B.V. Method and system for recommending products based on a ranking cocktail
US8543563B1 (en) 2012-05-24 2013-09-24 Xerox Corporation Domain adaptation for query translation
US9026425B2 (en) 2012-08-28 2015-05-05 Xerox Corporation Lexical and phrasal feature domain adaptation in statistical machine translation
US10452740B2 (en) 2012-09-14 2019-10-22 Sdl Netherlands B.V. External content libraries
US11386186B2 (en) 2012-09-14 2022-07-12 Sdl Netherlands B.V. External content library connector systems and methods
US11308528B2 (en) 2012-09-14 2022-04-19 Sdl Netherlands B.V. Blueprinting of multimedia assets
US9916306B2 (en) 2012-10-19 2018-03-13 Sdl Inc. Statistical linguistic analysis of source content
US9152622B2 (en) 2012-11-26 2015-10-06 Language Weaver, Inc. Personalized machine translation via online adaptation
US9235567B2 (en) 2013-01-14 2016-01-12 Xerox Corporation Multi-domain machine translation model adaptation
US9047274B2 (en) 2013-01-21 2015-06-02 Xerox Corporation Machine translation-driven authoring system and method
US9213694B2 (en) 2013-10-10 2015-12-15 Language Weaver, Inc. Efficient online domain adaptation
US9582499B2 (en) 2014-04-14 2017-02-28 Xerox Corporation Retrieval of domain relevant phrase tables
US9606988B2 (en) 2014-11-04 2017-03-28 Xerox Corporation Predicting the quality of automatic translation of an entire document
US9442922B2 (en) * 2014-11-18 2016-09-13 Xerox Corporation System and method for incrementally updating a reordering model for a statistical machine translation system
US9367541B1 (en) 2015-01-20 2016-06-14 Xerox Corporation Terminological adaptation of statistical machine translation system through automatic generation of phrasal contexts for bilingual terms
US10025779B2 (en) 2015-08-13 2018-07-17 Xerox Corporation System and method for predicting an optimal machine translation system for a user based on an updated user profile
US9836453B2 (en) 2015-08-27 2017-12-05 Conduent Business Services, Llc Document-specific gazetteers for named entity recognition
US10614167B2 (en) 2015-10-30 2020-04-07 Sdl Plc Translation review workflow systems and methods
US10635863B2 (en) 2017-10-30 2020-04-28 Sdl Inc. Fragment recall and adaptive automated translation
US10817676B2 (en) 2017-12-27 2020-10-27 Sdl Inc. Intelligent routing services and systems
US10747962B1 (en) 2018-03-12 2020-08-18 Amazon Technologies, Inc. Artificial intelligence system using phrase tables to evaluate and improve neural network based machine translation
US10769307B2 (en) 2018-05-30 2020-09-08 Bank Of America Corporation Processing system using natural language processing for performing dataset filtering and sanitization
US11256867B2 (en) 2018-10-09 2022-02-22 Sdl Inc. Systems and methods of machine learning for digital assets and message creation

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS57201958A (en) * 1981-06-05 1982-12-10 Hitachi Ltd Device and method for interpretation between natural languages
JPS6140672A (en) * 1984-07-31 1986-02-26 Hitachi Ltd Processing system for dissolution of many parts of speech
JP3345763B2 (en) 1994-03-04 2002-11-18 日本電信電話株式会社 Natural language translator
JPH1011447A (en) 1996-06-21 1998-01-16 Ibm Japan Ltd Translation method and system based upon pattern
JP3430007B2 (en) 1998-03-20 2003-07-28 富士通株式会社 Machine translation device and recording medium

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
FRANZ JOSEF OCH, HERMANN NEY: "Improved Statistical Alignment Models" ACL00: PROC. OF THE 38TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, [Online] 2 - 6 October 2000, pages 440-447, XP002279144 Hong Kong, China Retrieved from the Internet: <URL:http://www-i6.informatik.rwth-aachen. de/Colleagues/och/ACL00.ps> [retrieved on 2004-05-06] *
K. CHEN AND H. CHEN: "Machine Translation: An Integrated Approach" PROCEEDINGS OF THE SIXTH INTERNATIONAL CONFERENCE ON THEORETICAL AND METHODOLOGICAL ISSUES IN MACHINE TRANSLATION, [Online] 5 - 7 July 1995, pages 287-294, XP002279141 Leuven, Belgium *
P. KOEHN AND K. KNIGHT: "ChunkMT: Statistical Machine Translation with Richer Linguistic Knowledge" INFORMATION SCIENCES INSTITUTE, [Online] 20 April 2002 (2002-04-20), XP002279142 Marina del Rey, CA, USA Retrieved from the Internet: <URL:http://www.isi.edu/~koehn/publication s/chunkmt.pdf> [retrieved on 2004-05-06] *
TARO WATANABE, KENJI IMAMMURA AND EIICHIRO SUMITA: "Statistical Machine Translation Based on Hierarchical Phrase Alignment" PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON THEORETICAL AND METHODOLOGICAL ISSUES IN MACHINE TRANSLATION, [Online] 13 - 17 March 2002, XP002279140 Keihanna, Japan Retrieved from the Internet: <URL:http://www.eamt.org/archive/tmi2002/c onference/19_watanabe.pdf> [retrieved on 2004-05-06] *
VOGEL S ; OCH F -J ; NEY H: "The statistical translation module in the Verbmobil system" WORKSHOP ON MULTI-LINGUAL SPEECH COMMUNICATION, [Online] 11 - 13 October 2000, pages 69-74, XP002279143 Kyoto, Japan Retrieved from the Internet: <URL:http://www-i6.informatik.rwth-aachen. de/Colleagues/och/Translation_ATR_17Aug00. ps> [retrieved on 2004-05-06] *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8886517B2 (en) 2005-06-17 2014-11-11 Language Weaver, Inc. Trust scoring for language translation systems
US10319252B2 (en) 2005-11-09 2019-06-11 Sdl Inc. Language capability assessment and training apparatus and techniques
US8886518B1 (en) 2006-08-07 2014-11-11 Language Weaver, Inc. System and method for capitalizing machine translated text
US8990064B2 (en) 2009-07-28 2015-03-24 Language Weaver, Inc. Translating documents based on content
US10417646B2 (en) 2010-03-09 2019-09-17 Sdl Inc. Predicting the cost associated with translating textual content
US10984429B2 (en) 2010-03-09 2021-04-20 Sdl Inc. Systems and methods for translating textual content
US11003838B2 (en) 2011-04-18 2021-05-11 Sdl Inc. Systems and methods for monitoring post translation editing
US8942973B2 (en) 2012-03-09 2015-01-27 Language Weaver, Inc. Content page URL translation
US10261994B2 (en) 2012-05-25 2019-04-16 Sdl Inc. Method and system for automatic management of reputation of translators
US10402498B2 (en) 2012-05-25 2019-09-03 Sdl Inc. Method and system for automatic management of reputation of translators

Also Published As

Publication number Publication date
WO2003083709A3 (en) 2004-09-10
JP2005527894A (en) 2005-09-15
DE10392450T5 (en) 2005-04-21
CN1647068A (en) 2005-07-27
AU2003222126A1 (en) 2003-10-13
US20040024581A1 (en) 2004-02-05
US7624005B2 (en) 2009-11-24

Similar Documents

Publication Publication Date Title
US7624005B2 (en) Statistical machine translation
Durrani et al. A joint sequence translation model with integrated reordering
US8548794B2 (en) Statistical noun phrase translation
CA2408819C (en) Machine translation techniques
US9460080B2 (en) Modifying a tokenizer based on pseudo data for natural language processing
Specia et al. Predicting machine translation adequacy
Yang et al. Phrase-based backoff models for machine translation of highly inflected languages
CA2480398A1 (en) Phrase-based joint probability model for statistical machine translation
Watanabe et al. Left-to-right target generation for hierarchical phrase-based translation
Watanabe et al. Example-based decoding for statistical machine translation
Gildea Dependencies vs. constituents for tree-based alignment
Hayashi et al. Hierarchical phrase-based machine translation with word-based reordering model
Mrinalini et al. Pause-based phrase extraction and effective OOV handling for low-resource machine translation systems
JP5454763B2 (en) Device for associating words in a sentence pair and computer program therefor
Menezes et al. Syntactic models for structural word insertion and deletion during translation
Wang et al. Structure alignment using bilingual chunking
JP2006127405A (en) Method for carrying out alignment of bilingual parallel text and executable program in computer
Bisazza et al. Chunk-lattices for verb reordering in Arabic–English statistical machine translation: Special issues on machine translation for Arabic
Xiong et al. Linguistically annotated reordering: Evaluation and analysis
Sánchez-Martínez et al. Exploring the use of target-language information to train the part-of-speech tagger of machine translation systems
LEVEL Koehn et a].(45) Date of Patent: Nov. 24, 2009
Sánchez‐Martínez et al. Cooperative unsupervised training of the part-of-speech taggers in a bidirectional machine translation system
Dhekane Statistical approach with factored translation models for indian languages
Aghaebrahimian et al. The TransBank Aligner: Cross-Sentence Alignment with Deep Neural Networks
Mall et al. Shallow Parsing and Word Sense Disambiguation Used for Machine Translation from Hindi to English Languages.

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SC SD SE SG SK SL TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2003581063

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 20038070278

Country of ref document: CN

122 Ep: pct application non-entry in european phase