Machine Translation
Machine Translation
● Challenges
● Types
● Evaluation Metrics
● World Respective
● Evaluation Tracks
Overview - Machine Translation
● Machine Translation Tools
PRESENT ● References
FUTURE
Machine Translation (MT) is the application of
computers to the task of translating texts from one
natural language to another.
Google Translate:
Launched in April 2006 as
a statistical machine
translation service.
Source: https://en.wikipedia.org/wiki/Google_Translate
Ref. https://translate.google.co.in/
● Based on the availability of resources (corpus,
tools, speakers), there are two categories of
natural languages: high and low resource
languages.
● High resource languages are those languages
which are resource-rich languages like English,
German, French, and Hindi.
● Low resource languages are resource-poor like
many Indian languages especially found in the
north-eastern region of India like Boro, Khasi,
Kokborok, Mizo etc.
Is Machine Translation hard or easy?
● Lexical Ambiguity: Ambiguous words
have multiple, related or unrelated,
meanings.
■ Prepositions: in November
Challenges
Machine Translation
Word order of English (S-V-O) and Hindi
(S-O-V)
Word Proportion Example
order of languages languages
SOV 45% Proto-Indo-European, Sanskrit, Hindi, Ancient
Greek, Latin, Japanese
VSO 9% Biblical
Hebrew, Arabic, Irish, Filipino, Tuareg-Berber,
Welsh
https://en.wikipedia.org/wiki/Word_order
● Phrase Translation Problems:
● Semantic Ambiguity:
➢ Parallel data
➢ Thesaurus, WordNet, FrameNet, etc.
(Anoop, 2018)
● Rule-based: Relies on Set of Rules
■ Known as dictionary-based.
(Anoop, 2018)
● Interlingual-based Machine Translation:
(Anoop, 2018)
Drawback of Rule -based Machine Translation:
Machine Translation
● Example-based Machine Translation
(Anoop, 2018)
● Neural Machine Translation: Basics of
Modelling (seq-2-seq)
Types ■
■
Inputs: Parallel sentences
Source: https://opennmt.net/
● Neural Machine Translation: Basics of
Modelling (seq-2-seq)
Types
Machine Translation
https://www.researchgate.net/figure/Example-of-text-representation-by-one-hot-vector_fig2_301703031
● Embedding
Machine Translation
● Embedding
Types
Machine Translation
Source:
https://towardsdatascience.com/language-translation-with-rnns-d84d43b40571
Recurrent Neural Network (RNN) Model
https://towardsdatascience.com/understanding-encoder-decoder-sequence-to-sequence-model-679e04
Ref: https://www.youtube.com/watch?v=StOFwSRBwMo
af4346
Neural Machine Translation: Basic Recurrent
Neural Network (RNN) Model Disadvantage
■ Slow to train
Machine Translation
Source: https://www.youtube.com/watch?v=TQQlZhbC5ps
Neural Machine Translation: Recurrent Neural
Network (RNN) Model - LSTM
Types
Machine Translation
Embedding-based Encoder-Decoder Architecture
Source: https://towardsdatascience.com/language-translation-with-rnns-d84d43b40571
Neural Machine Translation: Recurrent Neural
Network (RNN) Model - Attention Mechanism
Types
Machine Translation
Attention Mechanism
Source: https://towardsdatascience.com/language-translation-with-rnns-d84d43b40571
Neural Machine Translation: Bidirectional
Recurrent Neural Network (BRNN) Model
[Sree Harsha Ramesh and Krishna Prasad Sankaranarayanan, 2018]
Types
Machine Translation
Source: https://images.app.goo.gl/KrDpfoWswawkPNka7
Disadvantage:
Types ●
●
Sequential process
Slower
Machine Translation
Types
Machine Translation
Source: https://ai.googleblog.com/2017/08/transformer-novel-neural-network.html
Neural Machine Translation: Transformer Model
Various Concept ●
●
Multi-modal Concept
Zero shot Concept
of MT
Neural Machine Translation: Multilingual Concept
[Tan et al., 2017]
Image source:
https://jlibovicky.github.io/2020/06/06/MT-Weekly-Unsupervised-Multimodal-MT.html
Neural Machine Translation: Zero-shot
Types
Machine Translation
Source: https://ai.googleblog.com/2016/11/zero-shot-translation-with-googles.html
● SMT v/s NMT: End to End Problem
Source He is a boy
Adequacy Fluency
Reference वह एक लड़का है
Machine Translation
Conference on Machine Translation (WMT)
Evaluation
■ http://www.statmt.org/
■ Hindi-Nepali, Hindi-Marathi,
Tracks ■
English-Tamil
2005-2020
Machine Translation
WMT Track Top Scorer Approach
Evaluation 2019
Nepali to NITS-CNLP NMT
Evaluation ■
■
http://lotus.kuee.kyoto-u.ac.jp/WAT/
English-Hindi, English-Odia,
Tracks Bengali/Hindi/Malayalam/Tamil/Telugu/
Marathi/Gujarati - English
■ 2014-2021
Machine Translation
WAT Track Top Scorer Approach
Evaluation English to
Hindi
Team: 683
(CNLP, NIT
Silchar)
NMT
(BRNN)
multimodal
Evaluation ■
■
English - Malayalam
English - Hindi
■ English - Punjabi
Tracks ●
■ 2017
Workshop on Low Resource Machine
Machine Translation Translation (LoResMT)
■ http://sites.google.com/view/loresmt/
■ Bhojpuri/Magahi/Sindhi-English,
Hindi-Bhojpuri, Hindi-Magahi,
Russian-Hindi
■ 2019-2021
● Conference On Machine Translation
World (WMT)
● NMT:
Available Tools OpenNMT: https://opennmt.net/
Machine Translation Marian: https://marian-nmt.github.io/
Nematus: https://github.com/EdinburghNLP/nematus
● Low Resource Indian Language
Translation
○ Mizo
Current Trends ○
○
Assamese
Nyshi
& ○ Khsai
Future Direction
● Multilingual-based Translation
Machine Translation
● Multimodal-based Translation
○ Hindi, Assamese
● Corpus creation and introduce in MT
Current Trends
& ● Efficiently tackle insufficient data issue
Future Direction
● Deals with:
Machine Translation ○ Out-of-Vocabulary
○ Rare-words
○ multi-word expressions problem
● Transfer learning:
Current Trends How can effectively utilize
&
● Text-to-Speech Translation
Future Direction
Machine Translation
● Speech-to-Speech Translation
References
1. Anoop Kunchukuttan, Pushpak Bhattacharyya. Faster Decoding for Subword Level Phrase-based SMT between Related Languages.
Proceedings of the Third Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial3), COLING, pp-82-88 (2016).
2. Sutskever, Ilya and Vinyals, Oriol and Le, Quoc V. Sequence to Sequence Learning with Neural Networks. In proceedings of the 27th
International Conference on Neural Information Processing Systems - volume 2, MIT Press, Cambridge, MA, USA, NIPS’14, pp.
3104–3112 (2014).
3. Sree Harsha Ramesh and Krishna Prasad Sankaranarayanan. Neural Machine Translation for Low Resource Languages using Bilingual
Lexicon Induced from Comparable Corpora. In proceedings of the 2018 Conference of the North American Chapter of the Association for
Computational Linguistics: Student Research Workshop. Association for Computational Linguistics, New Orleans, Louisiana, USA, pp.
112–119 (2018).
4. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention
is All you Need. In Advances in Neural Information Processing Systems 30, Curran Associates, Inc., pp. 5998–6008 (2017).
5. Tan, X., Chen, J., He, D., Xia, Y., Qin, T., Liu, T.Y. Multilingual neural machine translation with language clustering. In proceedings of the
2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language
Processing (EMNLP-IJCNLP), pp. 963–973. Association for Computational Linguistics, Hong Kong, China (2019).
6. Pulkit Madaan, Fatiha Sadat. Multilingual Neural Machine Translation involving Indian Languages. Proceedings of the WILDRE5– 5th
Workshop on Indian Language Data: Resources and Evaluation, European Language Resources Association (ELRA), pp. 29–32 (2020).
7. Sukanta Sen, Kamal Kumar Gupta, Asif Ekbal, Pushpak Bhattacharyya. IITP-MT at WAT2018: Transformer-based Multilingual
Indic-English Neural Machine Translation System, Association for Computational Linguistics (2018).
8. Calixto, I., Liu, Q., Campbell, N. Doubly-attentive decoder for multi-modal neural machine translation. In proceedings of the 55th Annual
Meeting of the Association for Computational Linguistics (Volume 1:Long Papers), pp. 1913–1924. Association for Computational
Linguistics, Vancouver, Canada (2017).
9. Koel Dutta Chowdhury, Mohammed Hasanuzzaman, Qun Liu. Multimodal Neural Machine Translation for Low-resource Language Pairs
using Synthetic Data. Proceedings of the Workshop on Deep Learning Approaches for Low-Resource NLP, Association for Computational
Linguistics, pp. 33-42 (2018).
10.
References
Bharathi Raja Chakravarthi, Ruba Priyadharshini, Bernardo Stearns, Arun Jayapal, Sridevy S, Mihael Arcan, Manel Zarrouk, John P McCrae. Multilingual
Multimodal Machine Translation for Dravidian Languages utilizing Phonetic Transcription. Proceedings of the 2nd Workshop on Technologies for MT of Low
Resource Languages. European Association for Machine Translation, pp-56-63 (2019).
11. Melvin Johnson, Mike Schuster, Quoc V. Le, Maxim Krikun, Yonghui Wu, Zhifeng Chen, Nikhil Thorat, Fernanda Viégas, Martin Wattenberg, Greg Corrado,
Macduff Hughes, Jeffrey Dean. Google’s Multilingual Neural Machine Translation System: Enabling Zero-Shot Translation. Transactions of the Association
for Computational Linguistics, Volume 5, pp-339–351 (2017).
12. Giulia MattoniPat NagleCarlos CollantesDimitar Sht. Shterionov Dimitar Sht. Shterionov. Zero-Shot Translation for Indian Languages with Sparse Data,
MT Summit (2017).
13. Rashi Kumar, Piyush Jha, Vineet Sahula. An Augmented Translation Technique for low Resource language pair: Sanskrit to Hindi translation. Proceedings
of the 2019 2nd International Conference on Algorithms, Computing and Artificial Intelligence, pp 377-383 (2019)
14. Saurav Kumar, Saunack Kumar, Diptesh Kanojia, Pushpak Bhattacharyya. A Passage to India”: Proceedings of the 1st Joint Workshop on Spoken Language
Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), pp. 352–357 (2020).
15. Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: A Method for Automatic Evaluation of Machine Translation. In proceedings
of the 40th Annual Meeting on Association for Computational Linguistics (ACL ’02). Association for Computational Linguistics, Stroudsburg, PA, USA,
pp-311–318.
16. Matthew Snover, Bonnie Dorr, Richard Schwartz, Linnea Micciulla, and John Makhoul. 2006. A study of translation edit rate with targeted human
annotation. InIn Proceedings of Association for Machine Translation in the Americas. pp. 223–231.
17. Alon Lavie and Michael J. Denkowski. 2009. The Meteor Metric for Automatic Evaluation of Machine Translation.Machine Translation 23, 2–3 (Sept.
2009), pp. 105–115.
18. Tomas Mikolov, Kai Chen, Greg Corrado and Jeffrey Dean. Efficient Estimation of Word Representations in Vector Space. 1st International Conference on
Learning Representations, {ICLR}, Scottsdale, Arizona, USA, May 2-4, 2013, Workshop Track Proceedings.
19. Jeffrey Pennington, Richard Socher, Christopher Manning. GloVe: Global Vectors for Word Representation. Proceedings of the 2014 Conference on
Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics, pp. 1532–1543.
20. Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language
Technologies, Volume 1, Association for Computational Linguistics, pp. 4171–4186.
Hands-on:
Machine Translation
Mr. Sahinur Laskar, PhD Scholar, NIT Silchar