IWSLT.2013

| Total: 42

#1 Human semantic MT evaluation with HMEANT for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Chi-kiu Lo ; Dekai Wu

We present the results of large-scale human semantic MT evaluation with HMEANT on the IWSLT 2013 German-English MT and SLT tracks and show that HMEANT evaluates the performance of the MT systems differently compared to BLEU and TER. Together with the references, all the translations are annotated by annotators who are native English speakers in both semantic role labeling stage and role filler alignment stage of HMEANT. We obtain high inter-annotator agreement and low annotation time costs which indicate that it is feasible to run a large-scale human semantic MT evaluation campaign using HMEANT. Our results also show that HMEANT is a robust and reliable semantic MT evaluation metric for running large-scale evaluation campaigns as it is inexpensive and simple while maintaining the semantic representational transparency to provide a perspective which is different from BLEU and TER in order to understand the performance of the state-of-the-art MT systems.

#2 English SLT and MT system description for the IWSLT 2013 evaluation [PDF] [Copy] [Kimi1] [REL]

Authors: Alexandra Birch ; Nadir Durrani ; Philipp Koehn

This paper gives a description of the University of Edinburgh’s (UEDIN) systems for IWSLT 2013. We participated in all the MT tracks and the German-to-English and Englishto-French SLT tracks. Our SLT submissions experimented with including ASR uncertainty into the decoding process via confusion networks, and looked at different ways of punctuating ASR output. Our MT submissions are mainly based on a system used in the recent evaluation campaign at the Workshop on Statistical Machine Translation [1]. We additionally explored the use of generalized representations (Brown clusters, POS and morphological tags) translating out of English into European languages.

#3 MSR-FBK IWSLT 2013 SLT system description [PDF] [Copy] [Kimi1] [REL]

Authors: Anthony Aue ; Qin Gao ; Hany Hassan ; Xiaodong He ; Gang Li ; Nicholas Ruiz ; Frank Seide

This paper describes the systems used for the MSR+FBK submission for the SLT track of IWSLT 2013. Starting from a baseline system we made a series of iterative and additive improvements, including a novel method for processing bilingual data used to train MT systems for use on ASR output. Our primary submission is a system combination of five individual systems, combining the output of multiple ASR engines with multiple MT techniques. There are two contrastive submissions to help place the combined system in context. We describe the systems used and present results on the test sets.

#4 Improving machine translation into Chinese by tuning against Chinese MEANT [PDF] [Copy] [Kimi1] [REL]

Authors: Chi-kiu Lo ; Meriem Beloucif ; Dekai Wu

We present the first ever results showing that Chinese MT output is significantly improved by tuning a MT system against a semantic frame based objective function, MEANT, rather than an n-gram based objective function, BLEU, as measured across commonly used metrics and different test sets. Recent work showed that by preserving the meaning of the translations as captured by semantic frames in the training process, MT systems for translating into English on both formal and informal genres are constrained to produce more adequate translations by making more accurate choices on lexical output and reordering rules. In this paper we describe our experiments in IWSLT 2013 TED talk MT tasks on tuning MT systems against MEANT for translating into Chinese and English respectively. We show that the Chinese translation output benefits more from tuning a MT system against MEANT than the English translation output due to the ambiguous nature of word boundaries in Chinese. Our encouraging results show that using MEANT is a promising alternative to BLEU in both evaluating and tuning MT systems to drive the progress of MT research across different languages.

#5 The NICT ASR system for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Chien-Lin Huang ; Paul R. Dixon ; Shigeki Matsuda ; Youzheng Wu ; Xugang Lu ; Masahiro Saiko ; Chiori Hori

This study presents the NICT automatic speech recognition (ASR) system submitted for the IWSLT 2013 ASR evaluation. We apply two types of acoustic features and three types of acoustic models to the NICT ASR system. Our system is comprised of six subsystems with different acoustic features and models. This study reports the individual results and fusion of systems and highlights the improvements made by our proposed methods that include the automatic segmentation of audio data, language model adaptation, speaker adaptive training of deep neural network models, and the NICT SprinTra decoder. Our experimental results indicated that our proposed methods offer good performance improvements on lecture speech recognition tasks. Our results denoted a 13.5% word error rate on the IWSLT 2013 ASR English test data set.

#6 FBK @ IWSLT 2013 – ASR tracks [PDF] [Copy] [Kimi1] [REL]

Authors: Daniele Falavigna ; Roberto Gretter ; Fabio Brugnara ; Diego Giuliani

This paper reports on the participation of FBK at the IWSLT2013 evaluation campaign on automatic speech recognition (ASR): precisely on both English and German ASR track. Only primary submissions have been sent for evaluation. For English, the ASR system features acoustic models trained on a portion of the TED talk recordings that was automatically selected according to the fidelity of the provided transcriptions. Two decoding steps are performed interleaved by acoustic feature normalization and acoustic model adaptation. A final step combines the outputs obtained after having rescored the word graphs generated in the second decoding step with 4 different language models. The latter are trained on: out-of-domain text data, in-domain data and several sets of automatically selected data. For German, acoustic models have been trained on automatically selected portions of a broadcast news corpus, called ”Euronews”. Differently from English, in this case only two decoding steps are carried out without making use of any rescoring procedure.

#7 QCRI at IWSLT 2013: experiments in Arabic-English and English-Arabic spoken language translation [PDF] [Copy] [Kimi1] [REL]

Authors: Hassan Sajjad ; Francisco Guzmán ; Preslav Nakov ; Ahmed Abdelali ; Kenton Murray ; Fahad Al Obaidli ; Stephan Vogel

We describe the Arabic-English and English-Arabic statistical machine translation systems developed by the Qatar Computing Research Institute for the IWSLT’2013 evaluation campaign on spoken language translation. We used one phrase-based and two hierarchical decoders, exploring various settings thereof. We further experimented with three domain adaptation methods, and with various Arabic word segmentation schemes. Combining the output of several systems yielded a gain of up to 3.4 BLEU points over the baseline. Here we also describe a specialized normalization scheme for evaluating Arabic output, which was adopted for the IWSLT’2013 evaluation campaign.

#8 A discriminative reordering parser for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Hwidong Na ; Jong-Hyeok Lee

We participated in the IWSLT 2013 Evaluation Campaign for the MT track for two official directions: German↔English. Our system consisted of a reordering module and a statistical machine translation (SMT) module under a pre-ordering SMT framework. We trained the reordering module using three scalable methods in order to utilize training instances as many as possible. The translation quality of our primary submissions were comparable to that of a hierarchical phrasebased SMT, which usually requires a longer time to decode.

#9 The RWTH Aachen machine translation systems for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Joern Wuebker ; Stephan Peitz ; Tamer Alkhouli ; Jan-Thorsten Peter ; Minwei Feng ; Markus Freitag ; Hermann Ney

This work describes the statistical machine translation (SMT) systems of RWTH Aachen University developed for the evaluation campaign International Workshop on Spoken Language Translation (IWSLT) 2013. We participated in the English→French, English↔German, Arabic→English, Chinese→English and Slovenian↔English MT tracks and the English→French and English→German SLT tracks. We apply phrase-based and hierarchical SMT decoders, which are augmented by state-of-the-art extensions. The novel techniques we experimentally evaluate include discriminative phrase training, a continuous space language model, a hierarchical reordering model, a word class language model, domain adaptation via data selection and system combination of standard and reverse order models. By application of these methods we can show considerable improvements over the respective baseline systems.

#10 Description of the UEDIN system for German ASR [PDF] [Copy] [Kimi1] [REL]

Authors: Joris Driesen ; Peter Bell ; Mark Sinclair ; Steve Renals

In this paper we describe the ASR system for German built at the University of Edinburgh (UEDIN) for the 2013 IWSLT evaluation campaign. For ASR, the major challenge to overcome, was to find suitable acoustic training data. Due to the lack of expertly transcribed acoustic speech data for German, acoustic model training had to be performed on publicly available data crawled from the internet. For evaluation, lack of a manual segmentation into utterances was handled in two different ways: by generating an automatic segmentation, and by treating entire input files as a single segment. Demonstrating the latter method is superior in the current task, we obtained a WER of 28.16% on the dev set and 36.21% on the test set.

#11 NTT-NAIST SMT systems for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Katsuhito Sudoh ; Graham Neubig ; Kevin Duh ; Hajime Tsukada

This paper presents NTT-NAIST SMT systems for English-German and German-English MT tasks of the IWSLT 2013 evaluation campaign. The systems are based on generalized minimum Bayes risk system combination of three SMT systems: forest-to-string, hierarchical phrase-based, phrasebased with pre-ordering. Individual SMT systems include data selection for domain adaptation, rescoring using recurrent neural net language models, interpolated language models, and compound word splitting (only for German-English).

#12 The 2013 KIT IWSLT speech-to-text systems for German and English [PDF] [Copy] [Kimi1] [REL]

Authors: Kevin Kilgour ; Christian Mohr ; Michael Heck ; Quoc Bao Nguyen ; Van Huy Nguyen ; Evgeniy Shin ; Igor Tseyzer ; Jonas Gehring ; Markus Müller ; Matthias Sperber ; Sebastian Stüker ; Alex Waibel

This paper describes our English Speech-to-Text (STT) systems for the 2013 IWSLT TED ASR track. The systems consist of multiple subsystems that are combinations of different front-ends, e.g. MVDR-MFCC based and lMel based ones, GMM and NN acoustic models and different phone sets. The outputs of the subsystems are combined via confusion network combination. Decoding is done in two stages, where the systems of the second stage are adapted in an unsupervised manner on the combination of the first stage outputs using VTLN, MLLR, and cMLLR.

#13 Polish-English speech statistical machine translation systems for the IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Krzysztof Wolk ; Krzysztof Marasek

This research explores the effects of various training settings from Polish to English Statistical Machine Translation system for spoken language. Various elements of the TED parallel text corpora for the IWSLT 2013 evaluation campaign were used as the basis for training of language models, and for development, tuning and testing of the translation system. The BLEU, NIST, METEOR and TER metrics were used to evaluate the effects of data preparations on translation results. Our experiments included systems, which use stems and morphological information on Polish words. We also conducted a deep analysis of provided Polish data as preparatory work for the automatic data correction and cleaning phase.

#14 The RWTH Aachen German and English LVCSR systems for IWSLT-2013 [PDF] [Copy] [Kimi1] [REL]

Authors: M. Ali Basha Shaik ; Zoltan Tüske ; Simon Wiesler ; Markus Nußbaum-Thom ; Stephan Peitz ; Ralf Schlüter ; Hermann Ney

In this paper, German and English large vocabulary continuous speech recognition (LVCSR) systems developed by the RWTH Aachen University for the IWSLT-2013 evaluation campaign are presented. Good improvements are obtained with state-of-the-art monolingual and multilingual bottleneck features. In addition, an open vocabulary approach using morphemic sub-lexical units is investigated along with the language model adaptation for the German LVCSR. For both the languages, competitive WERs are achieved using system combination.

#15 EU-BRIDGE MT: text translation of talks in the EU-BRIDGE project [PDF] [Copy] [Kimi1] [REL]

Authors: Markus Freitag ; Stephan Peitz ; Joern Wuebker ; Hermann Ney ; Nadir Durrani ; Matthias Huck ; Philipp Koehn ; Thanh-Le Ha ; Jan Niehues ; Mohammed Mediani ; Teresa Herrmann ; Alex Waibel ; Nicola Bertoldi ; Mauro Cettolo ; Marcello Federico

EU-BRIDGE1 is a European research project which is aimed at developing innovative speech translation technology. This paper describes one of the collaborative efforts within EUBRIDGE to further advance the state of the art in machine translation between two European language pairs, English→French and German→English. Four research institutions involved in the EU-BRIDGE project combined their individual machine translation systems and participated with a joint setup in the machine translation track of the evaluation campaign at the 2013 International Workshop on Spoken Language Translation (IWSLT). We present the methods and techniques to achieve high translation quality for text translation of talks which are applied at RWTH Aachen University, the University of Edinburgh, Karlsruhe Institute of Technology, and Fondazione Bruno Kessler. We then show how we have been able to considerably boost translation performance (as measured in terms of the metrics BLEU and TER) by means of system combination. The joint setups yield empirical gains of up to 1.4 points in BLEU and 2.8 points in TER on the IWSLT test sets compared to the best single systems.

#16 The MIT-LL/AFRL IWSLT-2013 MT system [PDF] [Copy] [Kimi1] [REL]

Authors: Michaeel Kazi ; Michael Coury ; Elizabeth Salesky ; Jessica Ray ; Wade Shen ; Terry Gleason ; Tim Anderson ; Grant Erdmann ; Lane Schwartz ; Brian Ore ; Raymond Slyh ; Jeremy Gwinnup ; Katherine Young ; Michael Hutt

This paper describes the MIT-LL/AFRL statistical MT system and the improvements that were developed during the IWSLT 2013 evaluation campaign [1]. As part of these efforts, we experimented with a number of extensions to the standard phrase-based model that improve performance on the Russian to English, Chinese to English, Arabic to English, and English to French TED-talk translation task. We also applied our existing ASR system to the TED-talk lecture ASR task. We discuss the architecture of the MIT-LL/AFRL MT system, improvements over our 2012 system, and experiments we ran during the IWSLT-2013 evaluation. Specifically, we focus on 1) cross-entropy filtering of MT training data, and 2) improved optimization techniques, 3) language modeling, and 4) approximation of out-of-vocabulary words.

#17 The speech recognition and machine translation system of IOIT for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Ngoc-Quan Pham ; Hai-Son Le ; Tat-Thang Vu ; Chi-Mai Luong

This paper describes the Automatic Speech Recognition (ASR) and Machine Translation (MT) systems developed by IOIT for the evaluation campaign of IWSLT2013. For the ASR task, using Kaldi toolkit, we developed the system based on weighted finite state transducer. The system is constructed by applying several techniques, notably, subspace Gaussian mixture models, speaker adaptation, discriminative training, system combination and SOUL, a neural network language model. The techniques used for automatic segmentation are also clarified. Besides, we compared different types of SOUL models in order to study the impact of words of previous sentences in predicting words in language modeling. For the MT task, the baseline system was built based on the open source toolkit N-code, then being augmented by using SOUL on top, i.e., in N-best rescoring phase.

#18 TÜBİTAK Turkish-English submissions for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Ertuğrul Yılmaz ; İlknur Durgar El-Kahlout ; Burak Aydın ; Zişan Sıla Özil ; Coşkun Mermer

This paper describes the TU ̈ B ̇ITAK Turkish-English submissions in both directions for the IWSLT’13 Evaluation Campaign TED Machine Translation (MT) track. We develop both phrase-based and hierarchical phrase-based statistical machine translation (SMT) systems based on Turkish wordand morpheme-level representations. We augment training data with content words extracted from itself and experiment with reverse word order for source languages. For the Turkish-to-English direction, we use Gigaword corpus as an additional language model with the training data. For the English-to-Turkish direction, we implemented a wide coverage Turkish word generator to generate words from the stem and morpheme sequences. Finally, we perform system combination of the different systems produced with different word alignments.

#19 FBK’s machine translation systems for the IWSLT 2013 evaluation campaign [PDF] [Copy] [Kimi1] [REL]

Authors: Nicola Bertoldi ; M. Amin Farajian ; Prashant Mathur ; Nicholas Ruiz ; Marcello Federico

This paper describes the systems submitted by FBK for the MT track of IWSLT 2013. We participated in the English-French as well as the bidirectional Persian-English translation tasks. We report substantial improvements in our English-French systems over last year’s baselines, largely due to improved techniques of combining translation and language models. For our Persian-English and English-Persian systems, we observe substantive improvements over baselines submitted by the workshop organizers, due to enhanced language-specific text normalization and the creation of a large monolingual news corpus in Persian.

#20 The Heidelberg University machine translation systems for IWSLT2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Patrick Simianer ; Laura Jehl ; Stefan Riezler

We present our systems for the machine translation evaluation campaign of the International Workshop on Spoken Language Translation (IWSLT) 2013. We submitted systems for three language directions: German-to-English, Russian-to-English and English-to-Russian. The focus of our approaches lies on effective usage of the in-domain parallel training data. Therefore, we use the training data to tune parameter weights for millions of sparse lexicalized features using efficient parallelized stochastic learning techniques. For German-to-English we incorporate syntax features. We combine all of our systems with large language models. For the systems involving Russian we also incorporate more data into building of the translation models.

#21 The UEDIN English ASR system for the IWSLT 2013 evaluation [PDF] [Copy] [Kimi1] [REL]

Authors: Peter Bell ; Fergus McInnes ; Siva Reddy Gangireddy ; Mark Sinclair ; Alexandra Birch ; Steve Renals

This paper describes the University of Edinburgh (UEDIN) English ASR system for the IWSLT 2013 Evaluation. Notable features of the system include deep neural network acoustic models in both tandem and hybrid configuration, cross-domain adaptation with multi-level adaptive networks, and the use of a recurrent neural network language model. Improvements to our system since the 2012 evaluation – which include the use of a significantly improved n-gram language model – result in a 19% relative WER reduction on the tst2012 set.

#22 The NAIST English speech recognition system for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Sakriani Sakti ; Keigo Kubo ; Graham Neubig ; Tomoki Toda ; Satoshi Nakamura

This paper describes the NAIST English speech recognition system for the IWSLT 2013 Evaluation Campaign. In particular, we participated in the ASR track of the IWSLT TED task. Last year, we participated in collaboration with Karlsruhe Institute of Technology (KIT). This year is our first time to build a full-fledged ASR system for IWSLT solely developed by NAIST. Our final system utilizes weighted finitestate transducers with four-gram language models. The hypothesis selection is based on the principle of system combination. On the IWSLT official test set our system introduced in this work achieves a WER of 9.1% for tst2011, 10.0% for tst2012, and 16.2% for the new tst2013.

#23 The KIT translation systems for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Than-Le Ha ; Teresa Herrmann ; Jan Niehues ; Mohammed Mediani ; Eunah Cho ; Yuqi Zhang ; Isabel Slawik ; Alex Waibel

In this paper, we present the KIT systems participating in all three official directions, namely English→German, German→English, and English→French, in translation tasks of the IWSLT 2013 machine translation evaluation. Additionally, we present the results for our submissions to the optional directions English→Chinese and English→Arabic. We used phrase-based translation systems to generate the translations. This year, we focused on adapting the systems towards ASR input. Furthermore, we investigated different reordering models as well as an extended discriminative word lexicon. Finally, we added a data selection approach for domain adaptation.

#24 The CASIA machine translation system for IWSLT 2013 [PDF] [Copy] [Kimi1] [REL]

Authors: Xingyuan Peng ; Xiaoyin Fu ; Wei Wei ; Zhenbiao Chen ; Wei Chen ; Bo Xu

In this paper, we describe the CASIA statistical machine translation (SMT) system for the IWSLT2013 Evaluation Campaign. We participated in the Chinese-English and English-Chinese translation tasks. For both of these tasks, we used a hierarchical phrase-based (HPB) decoder and made it as our baseline translation system. A number of techniques were proposed to deal with these translation tasks, including parallel sentence extraction, pre-processing, translation model (TM) optimization, language model (LM) interpolation, turning, and post-processing. With these techniques, the translation results were significantly improved compared with that of the baseline system.

#25 The AMARA corpus: building resources for translating the web’s educational content [PDF] [Copy] [Kimi1] [REL]

Authors: Francisco Guzman ; Hassan Sajjad ; Stephan Vogel ; Ahmed Abdelali

In this paper, we introduce a new parallel corpus of subtitles of educational videos: the AMARA corpus for online educational content. We crawl a multilingual collection community generated subtitles, and present the results of processing the Arabic–English portion of the data, which yields a parallel corpus of about 2.6M Arabic and 3.9M English words. We explore different approaches to align the segments, and extrinsically evaluate the resulting parallel corpus on the standard TED-talks tst-2010. We observe that the data can be successfully used for this task, and also observe an absolute improvement of 1.6 BLEU when it is used in combination with TED data. Finally, we analyze some of the specific challenges when translating the educational content.