Issue #57 – Simple and Effective Noisy Channel Modeling for Neural MT

24 Oct19

Issue #57 – Simple and Effective Noisy Channel Modeling for Neural MT

Author: Dr. Rohit Gupta, Sr. Machine Translation Scientist @ Iconic

Neural MT is widely used today and the results are undeniably better compared to the statistical machine translation (SMT) used earlier. One of the core components of an SMT system was the language model. In this post, we will look at how we can benefit from a language model in Neural MT, too. In particular, we will have a quick look at the paper from Yee et. al. (2019)

The probability modeling in SMT is indirect and called noisy channel modeling. In SMT, if our target is y and source is x, to compute target given source P(y|x), we compute source given target P(x|y) and P(y). Here, P(x|y) is called the translation model and P(y) is called the language model. In Neural MT, however, we directly predict P(y|x). We process the whole input sentence with encoder and feed it to the decoder which generates the target sentence. Now in Neural MT also, we can make use of probability scores from an external language model. However, to do so we
To finish reading, please visit source site

Leave a Reply