... 4.4 Translation time Since our splitting method is performed under left-to-right parsing, translation efficiency is not 426 Splitting Long or Ill-formed Input for Robust Spoken- languageTranslation ... sir for how many people please" Figure 3: Structure for (1) 3.2 Splitting input into well-formed parts and ill-formed parts Item (C) splits input into well-formed parts and ill-formed ... speech translation. In Proc. of ACL//EACL Workshop on Spoken Language Translation, pages 24-31. 427 to the right-neighboring string. If the whole in- put string can be covered with a passive...
... sentences tI1. Therefore, thesame techniques (translation models, decoder al-gorithm, etc.) which have been developed for SMT can be used in CAT.Note that the statistical models are defined atword ... framework, real-time is re-quired.4 Phrase-based models The usual statistical translationmodels can beclassified as single-word based alignment models. Models of this kind assume that an input word ... Association for Computational LinguisticsStatistical phrase-based modelsfor interactive computer-assisted translation Jes´us Tom´as and Francisco CasacubertaInstituto Tecnol´ogico de Inform´aticaUniversidad...
... Still, inverse IBM models perform better than the n-grams alone. Log-linear models show a bit of improvement with respect toIBM models. However, linear interpolated models perform the best. In ... because of poor language modelling that will affect on-line HTR decoding aswell. In fact, although language perplexities for thetest sets are quite low (33 for Spanish and 48 for En-glish), ... thesystem goal is not to produce “perfect” translationsin a completely automatic way, but to help the userbuild the translationwith the least effort possible.A typical approach to IMT is...
... then generated from a sin- gle formal language word using a translation model. The notion of what constitutes a natural clumping depends on the formal language. For example, sup- pose the English ... 1.17. Some trained translation probabilities are shown for the unigram and headword models in table 2. The formal language words have captured reason- able English words for their most likely ... alignment, with g(A) = g(C), and the ai denote the formal language word to which each e in c~ align. The individual words in a clump c are represented by el el(~). For all fertility models, ...
... has effectively used n-gram word sequence models as language models. Modern phrase-based translation using large scalen-gram languagemodels generally performs wellin terms of lexical choice, ... LMs. Syntactic language models have also been explored with tree-based translation models. Charniak et al. (2003) use syntactic lan-guage models to rescore the output of a tree-based translation ... machine translation viewed translation as a noisy channel process comprised ofa translation model, which functioned to posit ad-equate translations of source language words, anda target language...
... Translation with Local Language Models. In Proceedings of the2011 Conference on Empirical Methods in Natural Language Processing, pages 869–879, Edinburgh,Scotland, UK., July. Association for ... various hybrid LM variants. Translation quality is measured with BLEU, METEOR andTER (all in percentage form). The settings used for weight tuning are marked with †. Best models according toall ... Association for Computational Linguistics, pages 439–448,Avignon, France, April 23 - 27 2012.c2012 Association for Computational LinguisticsCutting the Long Tail: Hybrid Language Models for Translation...
... oppositeis true, with the LTM models achieving better per-formance.6On the NIST corpus, LTM-10 again achieves thebest gain of approximately 1 BLEU and up to 3 TER.LTM performs on par with or better ... 2 TER. Although theperformance on BLEU for both the 20 topic models LTM-20 and GTM-20 is suboptimal, the TER im-provement is better. Interestingly, the difference in translation quality between ... inducedsubdomains consistent within a document.Results Results for both settings are shown in Ta-ble 2. GTM models the latent topics at the documentlevel, while LTM models each sentence as a...
... Machine Translation The goal of the translation process in statisti-cal machine translation can be formulated as fol-lows: A source language stringis to be translated into a target language ... language. Those lexicon models lack from context infor-mation that can be extracted from the same paral-lel corpus. This additional information could be:Simple context information: information ofthe ... deal with this problem withinthe statistical framework is to use max-imum entropy methods. In this paper,we present how to use this type of in-formation within a statistical machinetranslation...
... 2001) for semantic rolelabeling. (Resnik et al., 2001) used backoff trans-lation lexicons for cross -language information re-trieval. More recently, (Xi and Hwa, 2005) haveused backoff modelsfor ... Nevertheless,standard SMT models tend to perform much bet-ter on languages that are morphologically simple,whereas highly inflected languages with a largenumber of potential word forms are more prob-lematic, ... Phrase Translation. Ph.D. the-sis, Information Sciences Institute, USC, Los Ange-les, California.P. Koehn. 2004. Pharaoh: a beam search decoder for phrase-based statistical machine translation models. In...
... creation of a translation forest from its rescoring with a lan-guage models or similar models. 3Since the struc-ture of the unied search space is context free (Đ3),we use the logic forlanguage ... gram-mar. For non-SCFG translation models, there aretwo kinds of edges. The first have zero tail nodes(i.e., an arity of 0), and correspond to word orphrase translation pairs (with all translation ... provides a training pipeline for discriminatively trained probabilistic translation models (Blunsom et al., 2008; Blunsom and Os-borne, 2008). In these models, the translation model is trained...
... number.ã For verbs, generated forms had to match theoriginal form for tense and negation.ã For adjectives, generated forms had to matchthe original form for degree of comparison andnegation.ã For ... The forms pro-duced by the tool from the lemma of an observed in-flected word form were subjected to several restric-tions:ã For nouns, generated forms had to match theoriginal form for number.ã ... modeling of inflected forms un-observed in training data and decoding proce-dures for a model with non-local target-sidefeature dependencies.1 Introduction Translation into languages with rich morphologypresents...
... language modelwe use for the experiments reported here is the sameas the one used for other experiments reported in thispaper.The results in Table 3 illustrate how the language model performs ... translation decodershave mostly relied on languagemodels to select theproper word order among many possible choices whentranslating between two languages. In this paper, weargue that a language ... limita-tion by replacing absolute word positions with relativepositions. The latter models define the distortion pa-rameters for a cept (one or more words). This models phrasal movement better since words...