neural machine paper 5
neural machine paper 5
Abstract: This article presents UnsupNMT, a 3-year project of which the first year
has already been completed. UnsupNMT proposes a radically different approach to
machine translation: unsupervised translation, that is, translation based on mono-
lingual data alone with no need for bilingual resources. This method is based on
deep learning of temporal sequences and uses cutting-edge interlingual word repre-
sentations in the form of cross-lingual word embeddings. This project is not only a
highly innovative proposal but it also opens a new paradigm in machine translation
which branches out to other disciplines, such us transfer learning. Despite the cu-
rrent limitations of unsupervised machine translation, the techniques developed are
expected to have great repercussions in areas where machine translation achieves
worse results, such as translation between languages which have little contact, e.g.
German and Russian.
Keywords: Machine Translation, Deep Learning, Word Embedding
Resumen: Este artı́culo presenta UnsupNMT, un proyecto de 3 años del que ha
trascurrido la primera anualidad. UnsupNMT plantea un método radicalmente dife-
rente de hacer traducción automática: la traducción no supervisada, es decir, basada
exclusivamente en textos monolingües sin ningún recurso bilingüe. El método pro-
puesto se basa en aprendizaje profundo de secuencias temporales combinado con los
últimos avances en representación interlingual de palabras (“cross-lingual word em-
beddings”). Además de ser una propuesta propiamente innovadora, abre un nuevo
paradigma de traducción automática con ramificaciones en otras disciplinas como
el aprendizaje por transferencia (“transfer learning”). A pesar de las limitaciones
actuales de la traducción automática no-supervisada, se espera que las técnicas desa-
rrolladas tengan gran repercusión en áreas donde la traducción automática consigue
peores resultados, como la traducción entre pares de idiomas con poco contacto,
tales como alemán o ruso.
Palabras clave: Traducción Automática, Aprendizaje Profundo, Word Embeddings
1 Introduction the dominant paradigm to machine transla-
tion (Bahdanau, Cho, and Bengio, 2014). As
Machine translation has been one of the most opposed to the traditional statistical machine
prominent applications of artificial intelligen- translation (SMT), NMT systems are trained
ce since the very beginnings of the field. In end-to-end, take advantage of continuous re-
addition to its intrinsic interest given the dif- presentations that greatly alleviate the spar-
ficulty and completeness of the problem, ma- sity problem, and make use of much larger
chine translation has a huge practical interest contexts, thus mitigating the locality pro-
in our increasingly global word, as it promises blem. Thanks to this, NMT has been repor-
to break the language barrier while keeping ted to significantly improve over SMT both
the cultural heritage and diversity of all the in automatic metrics and human evaluation.
languages spoken in the world.
In very recent times, previous approa- Current NMT methods require expensi-
ches have been superseded by neural machine ve annotated data, as they fail terribly when
translation (NMT), which has now become the training data is not big enough (Koehn
ISSN 1135-5948. DOI 10.26342/2019-63-18 © 2019 Sociedad Española para el Procesamiento del Lenguaje Natural
Mikel Artetxe, Gorka Labaka, Eneko Agirre