seq2 seq learning
TRANSCRIPT
Seq2Seq learningand recent advances
Vu Pham, 08.03.2017
Agenda
●
●
●
2
Recap: CNN for NLP
● →
●
●
●3
Recurrent cell
4
LSTM cell
5
Recurrent neural nets
6
C
●
●
Avoid explicit segmentation of the output sequence
●○○
●
7
Seq2Seq: intro
●
○ →
●
○ →
●
○
8
Seq2Seq: idea
9
I go to the zoo <EOS>
Je vais au zoo <EOS>
C
Encoder Decoder
Sequence to Sequence Learning with Neural Networks
Seq2Seq: Training
10
C
I go to the zoo <EOS>
Je vais au zoo <EOS>
Je vais au zoo
→ ∅ ∅ ∅ ∅ ∅
→ ∅ ∅ ∅ ∅ ∅
Seq2Seq: Inference
11
I go to the zoo <EOS>
Je vais
Je
au
vais
zoo
au
<EOS>
zoo
Seq2Seq: Inference with (pruned) beam search
12
Je
Elle
Ils
... <EOS>
allé
...
au
......
...
vais
suis
...
...
...
...
...
...
...
Attention mechanism
13
I go to the …... <EOS>
h0 h1 h2 h3 ... hn
c
MLPa0 a1 a2 a3
b0 b1 b2 b3
Je
Implemented in Tensorflow seq2seq()
Neural Machine Translation by Jointly Learning to Align and TranslateShow, Attend and Tell: Neural Image Caption Generation with Visual Attention
What we learned
●
●
●
14
Machine Translation
15
Image Captioning
16
VGG
14 x 14 x 512 196 x 512
Seq2Seq
196 sequences
Grammar parser
17
Grammar as a Foreign Language, NIPS 2014
Conversational bot
18
A Neural Conversational Model, ICML Workshop 2015
Skipthoughts
19
Skip-Thought Vectors, NIPS 2015
What else?●
○○○○○
●○○
●○○
20
What we learned
●
●
●○
○
○
○
21