DE eng

Search in the Catalogues and Directories

Page: 1 2 3
Hits 1 – 20 of 41

1
The Karlsruhe Institute of Technology Translation Systems for the WMT 2012
Herrmann, T.; Zhang, Y.; Waibel, A.. - : Association for Computational Linguistics, 2022
BASE
Show details
2
ELITR multilingual live subtitling: Demo and strategy
BASE
Show details
3
Multilingual Neural Translation
Ha, Thanh-Le [Verfasser]; Waibel, A. [Akademischer Betreuer]. - Karlsruhe : KIT-Bibliothek, 2020
DNB Subject Category Language
Show details
4
DaCToR: A data collection tool for the RELATER project ...
Hussain, J.; Zenkri, O.; Stüker, S.. - : European Language Resources Association (ELRA), 2020
BASE
Show details
5
Self-attentional models for lattice inputs
Sperber, M.; Waibel, A.; Neubig, G.. - : Association for Computational Linguistics, 2020
BASE
Show details
6
Multilingual Neural Translation
Ha, Thanh-Le. - : KIT-Bibliothek, Karlsruhe, 2020
Abstract: Machine translation (MT) refers to the technology that can automatically translate contents in one language into other languages. Being an important research area in the field of natural language processing, machine translation has typically been considered one of most challenging yet exciting problems. Thanks to research progress in the data-driven statistical machine translation (SMT), MT is recently capable of providing adequate translation services in many language directions and it has been widely deployed in various practical applications and scenarios. Nevertheless, there exist several drawbacks in the SMT framework. The major drawbacks of SMT lie in its dependency in separate components, its simple modeling approach, and the ignorance of global context in the translation process. Those inherent drawbacks prevent the over-tuned SMT models to gain any noticeable improvements over its horizon. Furthermore, SMT is unable to formulate a multilingual approach in which more than two languages are involved. The typical workaround is to develop multiple pair-wise SMT systems and connect them in a complex bundle to perform multilingual translation. Those limitations have called out for innovative approaches to address them effectively. On the other hand, it is noticeable how research on artificial neural networks has progressed rapidly since the beginning of the last decade, thanks to the improvement in computation, i.e faster hardware. Among other machine learning approaches, neural networks are known to be able to capture complex dependencies and learn latent representations. Naturally, it is tempting to apply neural networks in machine translation. First attempts revolve around replacing SMT sub-components by the neural counterparts. Later attempts are more revolutionary by fundamentally changing the whole core of SMT with neural networks, which is now popularly known as neural machine translation (NMT). NMT is an end-to-end system which directly estimate the translation model between the source and target sentences. Furthermore, it is later discovered to capture the inherent hierarchical structure of natural language. This is the key property of NMT that enables a new training paradigm and a less complex approach for multilingual machine translation using neural models. This thesis plays an important role in the evolutional course of machine translation by contributing to the transition of using neural components in SMT to the completely end-to-end NMT and most importantly being the first of the pioneers in building a neural multilingual translation system. First, we proposed an advanced neural-based component: the neural network discriminative word lexicon, which provides a global coverage for the source sentence during the translation process. We aim to alleviate the problems of phrase-based SMT models that are caused by the way how phrase-pair likelihoods are estimated. Such models are unable to gather information from beyond the phrase boundaries. In contrast, our discriminative word lexicon facilitates both the local and global contexts of the source sentences and models the translation using deep neural architectures. Our model has improved the translation quality greatly when being applied in different translation tasks. Moreover, our proposed model has motivated the development of end-to-end NMT architectures later, where both of the source and target sentences are represented with deep neural networks. The second and also the most significant contribution of this thesis is the idea of extending an NMT system to a multilingual neural translation framework without modifying its architecture. Based on the ability of deep neural networks to modeling complex relationships and structures, we utilize NMT to learn and share the cross-lingual information to benefit all translation directions. In order to achieve that purpose, we present two steps: first in incorporating language information into training corpora so that the NMT learns a common semantic space across languages and then force the NMT to translate into the desired target languages. The compelling aspect of the approach compared to other multilingual methods, however, lies in the fact that our multilingual extension is conducted in the preprocessing phase, thus, no change needs to be done inside the NMT architecture. Our proposed method, a universal approach for multilingual MT, enables a seamless coupling with any NMT architecture, thus makes the multilingual expansion to the NMT systems effortlessly. Our experiments and the studies from others have successfully employed our approach with numerous different NMT architectures and show the universality of the approach. Our multilingual neural machine translation accommodates cross-lingual information in a learned common semantic space to improve altogether every translation direction. It is then effectively applied and evaluated in various scenarios. We develop a multilingual translation system that relies on both source and target data to boost up the quality of a single translation direction. Another system could be deployed as a multilingual translation system that only requires being trained once using a multilingual corpus but is able to translate between many languages simultaneously and the delivered quality is more favorable than many translation systems trained separately. Such a system able to learn from large corpora of well-resourced languages, such as English → German or English → French, has proved to enhance other translation direction of low-resourced language pairs like English → Lithuania or German → Romanian. Even more, we show that kind of approach can be applied to the extreme case of zero-resourced translation where no parallel data is available for training without the need of pivot techniques. The research topics of this thesis are not limited to broadening application scopes of our multilingual approach but we also focus on improving its efficiency in practice. Our multilingual models have been further improved to adequately address the multilingual systems whose number of languages is large. The proposed strategies demonstrate that they are effective at achieving better performance in multi-way translation scenarios with greatly reduced training time. Beyond academic evaluations, we could deploy the multilingual ideas in the lecture-themed spontaneous speech translation service (Lecture Translator) at KIT. Interestingly, a derivative product of our systems, the multilingual word embedding corpus available in a dozen of languages, can serve as a useful resource for cross-lingual applications such as cross-lingual document classification, information retrieval, textual entailment or question answering. Detailed analysis shows excellent performance with regard to semantic similarity metrics when using the embeddings on standard cross-lingual classification tasks.
Keyword: DATA processing & computer science; ddc:004; info:eu-repo/classification/ddc/004
URL: https://doi.org/10.5445/IR/1000104498
https://publikationen.bibliothek.kit.edu/1000104498/57353596
https://publikationen.bibliothek.kit.edu/1000104498
BASE
Hide details
7
DaCToR: A data collection tool for the RELATER project
Hussain, J.; Zenkri, O.; Stüker, S.. - : European Language Resources Association, 2020
BASE
Show details
8
Paraphrases as foreign languages in multilingual neural machine translation
Waibel, A.; Zhou, Z.; Sperber, M.. - : Association for Computational Linguistics, 2019
BASE
Show details
9
Neural language codes for multilingual acoustic models
In: ISSN: 2308-457X (2018)
BASE
Show details
10
Multilingual Modulation by Neural Language Codes
Müller, Markus. - : KIT-Bibliothek, Karlsruhe, 2018
BASE
Show details
11
Learning from Noisy Data in Statistical Machine Translation
Mediani, Mohammed. - : KIT-Bibliothek, Karlsruhe, 2017
BASE
Show details
12
Machine Translation of Spontaneous Speech
Cho, Eunah [Verfasser]; Waibel, A. [Akademischer Betreuer]. - Karlsruhe : KIT-Bibliothek, 2016
DNB Subject Category Language
Show details
13
Linguistic Structure in Statistical Machine Translation
Herrmann, Teresa [Verfasser]; Waibel, A. [Akademischer Betreuer]. - Karlsruhe : KIT-Bibliothek, 2015
DNB Subject Category Language
Show details
14
Online Incremental Machine Translation
Rottmann, Kay [Verfasser]; Waibel, A. [Akademischer Betreuer]. - Karlsruhe : KIT-Bibliothek, 2015
DNB Subject Category Language
Online dissertations
Show details
15
Linguistic Structure in Statistical Machine Translation
Herrmann, Teresa. - : KIT-Bibliothek, Karlsruhe, 2015
BASE
Show details
16
Multilingual shifting deep bottleneck features for low-resource ASR
Nguyen, Q. B.; Gehring, J.; Mueller, M.. - : Institute of Electrical and Electronics Engineers, 2014
BASE
Show details
17
Training time reduction and performance improvements from multilingual techniques on the BABEL ASR task
Stuker, S.; Muller, M.; Nguyen, Q. B.. - : Institute of Electrical and Electronics Engineers, 2014
BASE
Show details
18
An MT Error-driven Discriminative Word Lexicon using Sentence Structure Features
Niehues, J.; Waibel, A.. - : Curran Associates, Inc., 2013
BASE
Show details
19
Combining Word Reordering Methods on different Linguistic Abstraction Levels for Statistical Machine Translation
Niehues, J.; Herrmann, T.; Waibel, A.. - : Association for Computational Linguistics, 2013
BASE
Show details
20
Parallel Phrase Scoring for Extra-large Corpora
In: The Prague bulletin of mathematical linguistics, 98 (1), 87-98 ; ISSN: 0032-6585, 1804-0462 (2012)
BASE
Show details

Page: 1 2 3

Catalogues
0
0
4
0
8
0
0
Bibliographies
1
0
0
0
0
0
0
0
3
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
26
0
0
4
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern