Neural Machine Translation (NMT) uses neural networks to train a framework to translate from one language to another. The semantics of languages differ so starkly that it becomes challenging to prescribe a phrase-to-phrase or word-to-word translation system.
Note: Using neural networks to train a translation system decreases reliance on individual words and instead focuses on the larger context of the sentence
Largely, NMT systems tend to contain a series of deep neural networks that can be trained to translate an entire sentence, paragraph or even a document.
Generally, the input sequence is encoded into a sequence of numbers. The resultant sequence of numbers is eventually fed into the decoding network that converts the sequence into the output sequence of words.
This approach is generally known as the Cho encoder-decoder framework.
The input sequence,
Consequently, as per the
NMT systems that comply with the above probabilistic model are known as L2R auto-regressive compliant.
Each linguistic token in the input sequence is discretely converted into a vector of its own.
The encoder uses an RNN to model the dependent patterns embedded within the input sequence. The corresponding encoding RNN model can be represented as the following:
The resultant encoded sequence,
The decoding framework also uses RNN using the initial state being simply the encoded sequence,
The final decoded sequence can be represented as
The decoded state,
Index
On the whole, neural machine translation provides a robust framework that forms a one-stop solution for much of the language-translation problems that especially fail to capture the contextual cues. It is widely accepted and used in popular services such as the Google Neural Machine Translation System.
Free Resources