|
|
(One intermediate revision by the same user not shown) |
Line 1: |
Line 1: |
| =Introduction=
| |
|
| |
|
| ==Background Knowledge==
| |
| *NTM
| |
| '''Neural Machine Translation (NMT)''', which is based on deep neural networks and provides an end- to-end solution to machine translation, uses an '''RNN-based encoder-decoder architecture''' to model the entire translation process. Specifically, an NMT system first reads the source sentence using an encoder to build a "thought" vector, a sequence of numbers that represents the sentence meaning; a decoder, then, processes the "meaning" vector to emit a translation. (Figure 1)<sup>[[#References|[1]]]</sup>
| |
| [[File:VNFigure1.png|thumb|600px|center|Figure 1: Encoder-decoder architecture – example of a general approach for NMT.]]
| |
|
| |
| *Sequence-to-Sequence(Seq2Seq) Model
| |
| [[File:VNFigure4.png|thumb|500px|center|Figure 2: Seq2Seq Model]]
| |
| - Two RNNs: an encoder RNN, and a decoder RNN
| |
| 1) The input is passed though the encoder and it’s final hidden state, the “thought vector” is passed to the decoder as it’s initial hidden state.
| |
| 2)Decoder given the start of sequence token, <SOS>, and iteratively produces output until it outputs the end of sequence token, <EOS>
| |
| - Commonly used in text generation, machine translation, and related problems
| |
|
| |
| *Beam Search
| |
| Decoding process:
| |
| [[File:VNFigure2.png|thumb|600px|center|Figure 3]]
| |
| Problem: Choosing the word with highest score at each time step t is not necessarily going to give you the sentence with the highest probability(Figure 3). Beam search solves this problem (Figure 4). Beam search has a size m such that at each time step t, it takes the top m proposal and continues decoding with each one of them. In the end, you will get a sentence with the highest probability not in the word level.
| |
| [[File:VNFigure3.png|thumb|600px|center|Figure 4]]
| |
|
| |
| =References=
| |
|
| |
| 1. https://github.com/tensorflow/nmt
| |