part2:deep learning - hitir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-deep_learning.pdf ·...

47
Part 2: Deep Learning 2017-11-27 IJCNLP 2017 Tutorial 1

Upload: others

Post on 21-May-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Part 2: DeepLearning

2017-11-27 IJCNLP 2017 Tutorial 1

Page 2: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Part 2.1: DeepLearning Background

2017-11-27 IJCNLP 2017 Tutorial 2

Page 3: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

WhatisMachineLearning?

• FromDatatoKnowledge

InputAlgorithm

Output

Traditional Program

InputOutput

“Algorithm”

ML Program2017-11-27 IJCNLP 2017 Tutorial 3

Page 4: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

AStandardExampleofML

• TheMNIST(ModifiedNIST)databaseofhand-writtendigitsrecognition– Publiclyavailable– AhugeamountabouthowwellvariousMLmethodsdoonit

– 60,000+ 10,000 hand-writtendigits(28x28pixelseach)

2017-11-27 IJCNLP 2017 Tutorial 4

Page 5: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Veryhardtosaywhatmakesa2

2017-11-27 IJCNLP 2017 Tutorial 5

Page 6: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Traditional Model(before2012)

• Fixed/engineeredfeatures+trainableclassifier– Designingafeatureextractorrequiresconsiderableeffortsbyexperts

SIFTGISTShapecontext

2017-11-27 IJCNLP 2017 Tutorial 6

Page 7: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DeepLearning (after2012)

• LearningHierarchicalRepresentations• DEEPmeansmorethanonestageofnon-linearfeaturetransformation

2017-11-27 IJCNLP 2017 Tutorial 7

Page 8: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DeepLearningArchitecture

2017-11-27 IJCNLP 2017 Tutorial 8

Page 9: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DeepLearningisNotNew

• 1980stechnology(NeuralNetworks)

2017-11-27 IJCNLP 2017 Tutorial 9

Page 10: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

AboutNeuralNetworks

• Pros– Simpletolearnp(y|x)– Performance is OKforshallownets

• Cons– Troublewith>3layers– Overfitts– Slowtotrain

2017-11-27 IJCNLP 2017 Tutorial 10

Page 11: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DeepLearningbeatsNN

• Pros– Simpletolearnp(y|x)

– Performance is OK forshallownets

• Cons– Troublewith>3layers

– Overfitts

– Slowtotrain

• Dropout• Maxout• StochasticPooling

GPU

• Newactivationfunctions:ReLU,…

• Gatedmechanism

2017-11-27 IJCNLP 2017 Tutorial 11

Page 12: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

ResultsonMNIST

• NaïveNeuralNetwork– 96.59%

• SVM(defaultsettingsforlibsvm)– 94.35%

• OptimalSVM[AndreasMueller]– 98.56%

• Thestateoftheart:ConvolutionalNN(2013)– 99.79%

2017-11-27 IJCNLP 2017 Tutorial 12

Page 13: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DeepLearningforSpeechRecognition

2017-11-27 IJCNLP 2017 Tutorial 13

Page 14: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DL forNLP: Representation Learning

A

B

A

B

A

B

- -

---

-

2017-11-27 IJCNLP 2017 Tutorial 14

Page 15: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DL forNLP: End-to-End Learning

2017-11-27 IJCNLP 2017 Tutorial 15

ROOThas_VBZ good_JJ Control_NN ._.

Stack Buffer

He_PRPnsubj

Configuration

0 0 1 0 1 … 0 1 0 0

Traditional Parser Stack-LSTM Parser

Page 16: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Part 2.2: Feedforward Neural Networks

2017-11-27 IJCNLP 2017 Tutorial 16

Page 17: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

feature units

decision unit

input units

hand-coded programs

learned weights

TheStandardPerceptronArchitecture

IPhone is very good .

good very/good very …

>5?

0.8 0.9 0.1 …

2017-11-27 IJCNLP 2017 Tutorial 17

Page 18: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

TheLimitationsofPerceptrons

• Thehand-codedfeatures– Greatinfluenceontheperformance– Needlotsofcosttofindsuitablefeatures

• Alinearclassifierwithahyperplane– Cannotseparatenon-lineardata,suchasXOR functioncannotbelearnedbyasingle-layerperceptron

0,1

0,0 1,0

1,1

weight plane output =1output =0

The positive and negative casescannot be separated by a plane

2017-11-27 IJCNLP 2017 Tutorial 18

Page 19: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

LearningwithNon-linearHiddenLayers

2017-11-27 IJCNLP 2017 Tutorial 19

Page 20: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

FeedforwardNeuralNetworks

• Multi-layerPerceptron(MLP)• Theinformationispropagatedfrom

theinputstotheoutputs• NOcyclebetweenoutputsandinputs• Learningtheweightsofhiddenunits

isequivalenttolearningfeatures• Networkswithouthiddenlayersare

verylimitedintheinput-outputmappings– Morelayersoflinearunitsdonot

help.Itsstilllinear– Fixedoutputnon-linearities arenot

enoughx1 x2 xn…..

1st hidden layer

2nd hiddenlayer

Output layer

2017-11-27 IJCNLP 2017 Tutorial 20

Page 21: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

MultipleLayer NeuralNetworks

• Whatarethosehiddenneuronsdoing?– Mayberepresentoutlines

2017-11-27 IJCNLP 2017 Tutorial 21

Page 22: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

GeneralOptimizing(Learning)Algorithms

• GradientDescent

• StochasticGradientDescent(SGD)– Minibatch SGD(m>1),OnlineGD(m=1)

2017-11-27 IJCNLP 2017 Tutorial 22

Page 23: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Computational/Flow Graphs

• DescribingMathematicalExpressions• Forexample

– e=(a+b)*(b+1)• c=a+b,d=b+1,e=c*d

– Ifa=2,b=1

2017-11-27 IJCNLP 2017 Tutorial 23

Page 24: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

DerivativesonComputationalGraphs

2017-11-27 IJCNLP 2017 Tutorial 24

Page 25: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

ComputationalGraphBackwardPass(Backpropagation)

2017-11-27 IJCNLP 2017 Tutorial 25

Page 26: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Part 2.3: Recurrent and OtherNeural Networks

2017-11-27 IJCNLP 2017 Tutorial 26

Page 27: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

LanguageModels

• Alanguagemodelcomputesaprobabilityforasequenceofword:!(#$,⋯#') orpredictsaprobabilityforthenextword:!(#')$|#$,⋯#')

• Usefulformachinetranslation,speechrecognition,andsoon– Wordordering

• P(thecatissmall)>P(smalltheiscat)

– Wordchoice• P(therearefour cats)>P(therearefor cats)

2017-11-27 IJCNLP 2017 Tutorial 27

Page 28: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

TraditionalLanguageModels

• AnincorrectbutnecessaryMarkovassumption!– Probabilityisusuallyconditionedonn previouswords– ! "#,⋯"& =∏ !("*|"#,⋯ ,"*,#) ≈/*0# ∏ !("*|"*,(&,#),⋯ ,"*,#)/

*0#• Disadvantages

– ThereareALOTofn-grams!– Cannotseetoolonghistory

2017-11-27 IJCNLP 2017 Tutorial 28

Page 29: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

RecurrentNeuralNetworks(RNNs)

• Conditiontheneuralnetworkonallpreviousinputs• RAMrequirementonlyscaleswithnumberofinputs

W1 W1

ht ht+1ht-1

yt-1 yt yt+1

xt-1 xt xt+1

W1

W2 W2W2

W3 W3W3

2017-11-27 IJCNLP 2017 Tutorial 29

Page 30: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

RecurrentNeuralNetworks (RNNs)

• Atasingletimestept– ℎ" = tanh()*ℎ"+* +)-.")– 01" = 234567.()8ℎ")

ht …ht-1

01"

xt

W1

W2

W3

ht

01"

xt

W1

W2

W3

W1

2017-11-27 IJCNLP 2017 Tutorial 30

Page 31: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

TrainingRNNsishard

• Ideallyinputsfrommanytimestepsagocanmodifyoutputy• Forexample,with2timesteps

ht ht+1ht-1

yt-1 yt yt+1

xt-1 xt xt+1

W1 W1

W2 W2

W3 W3

2017-11-27 IJCNLP 2017 Tutorial 31

Page 32: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

BackPropagation ThroughTime(BPTT)• Totalerroristhesumofeacherrorattimestept

– !"!# = ∑ !"&

!#'()*

• !"&!#+ =

!"&!,&

!,&!#+ iseasytobecalculated

• Buttocalculate !"&!#. =!"&!,&

!,&!/&

!/&!#. ishard(alsofor01)

• Becauseℎ( = tanh(0*ℎ(8* +01:() dependsonℎ(8*,whichdependson0* andℎ(81,andsoon.

• So !"&!#. = ∑ !"&!,&

!,&!/&

!/&!/<

!/<!#.

(=)*

ht

>?(

xt

W1

W2

W3

2017-11-27 IJCNLP 2017 Tutorial 32

Page 33: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

The vanishinggradientproblem• !"#

!$= ∑ !"#

!'#

!'#!(#

!(#!()

!()!$

*+,- ,ℎ* = tanh(4-ℎ*5- +4

78*)

• !(#!()

= ∏!(;!(;<=

*>,+?- = ∏ 4-diag[tanh′(⋯ )]*

>,+?-

• !(;!(;<=

≤ H 4- ≤ HI-– where H is bound diag[tanh′(⋯ )] , I-is the largest singular value of4-

• !(#!()

≤ HI-*5+à0

– ifHI- < 1, thiscanbecomeverysmall (vanishinggradient)– ifHI- > 1, thiscanbecomeverylarge(explodinggradient)

• Trickforexplodinggradient:clippingtrick(setathreshold)

2017-11-27 IJCNLP 2017 Tutorial 33

Page 34: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

A“solution”

• Intuition– Ensure!"# ≥ 1à topreventvanishinggradients

• So…– ProperinitializationoftheW– TouseReLU insteadoftanh orsigmoidactivationfunctions

2017-11-27 IJCNLP 2017 Tutorial 34

Page 35: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Abetter“solution”

• Recalltheoriginaltransitionequation– ℎ" = tanh()*ℎ"+* +)-.")

• Wecaninsteadupdatethestateadditively– 0" = tanh()*ℎ"+* +)-.")– ℎ" = ℎ"+* + 0"– then, 123

12345= 1 + 173

12345≥ 1

– Ontheotherhand• ℎ" = ℎ"+* + 0" = ℎ"+- + 0"+* + 0" = ⋯

ht …ht-1

:;"

xt

2017-11-27 IJCNLP 2017 Tutorial 35

Page 36: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Abetter“solution” (cont.)

• Interpolatebetweenoldstateandnewstate(“choosingtoforget”)– !" = $ %&'" + )&ℎ"+,– ℎ" = !" ⊙ ℎ"+, + (1 − !") ⊙ 2"

• Introduceaseparateinputgate3"– 3" = $ %4'" + )4ℎ"+,– ℎ" = !" ⊙ ℎ"+, + 3" ⊙ 2"

• Selectivelyexposememorycell5" withanoutputgate6"– 6" = $ %7'" + )7ℎ"+,– 5" = !" ⊙ 5"+, + 3" ⊙ 2"– ℎ" = 6" ⊙ tanh(5")2017-11-27 IJCNLP 2017 Tutorial 36

Page 37: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

LongShort-TermMemory(LSTM)

• Hochreiter & Schmidhuber,1997• LSTM=additiveupdates+gating

Xt

+ + + +0 1 2 3

ht-1

Ct-1 Ct

ht

+

�σ σ σtanh

tanh

ht

2017-11-27 IJCNLP 2017 Tutorial 37

Page 38: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

GatedRecurrentUnites,GRU(Choetal.2014)

• Mainideas– Keeparoundmemoriestocapturelongdistancedependencies– Allowerrormessagestoflowatdifferentstrengthsdependingontheinputs

• Updategate– Basedoncurrentinputandhiddenstate– !" = $ %&'" + )&ℎ"+,

• Resetgate– Similarlybutwithdifferentweights– -" = $(%/'" + )/ℎ"+,)

2017-11-27 IJCNLP 2017 Tutorial 38

Page 39: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

GRU• Memoryattimestepcombinescurrentandprevioustimesteps

– ℎ" = $" ⊙ ℎ"&' + (1 − $") ⊙ ℎ-– Updategatez controlshowmuchofpaststateshouldmatternow

• Ifz closedto1,thenwecancopyinformationinthatunitthroughmanytimestepsà lessvanishinggradient!

• Newmemorycontent– ℎ-" = tanh(23" + 4" ⊙ 5ℎ"&')– Ifresetgater unitiscloseto0,thenthisignorespreviousmemoryandonlystoresthenewinputinformationà allowsmodeltodropinformationthatisirrelevantinthefuture

2017-11-27 IJCNLP 2017 Tutorial 39

Page 40: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

LSTMvs.GRU

• Noclearwinner!• Tuninghyperparameters likelayersizeisprobablymoreimportantthanpickingtheidealarchitecture

• GRUshave fewer parametersandthusmaytrainabitfasterorneedless datatogeneralize

• Ifyouhaveenoughdata,thegreaterexpressivepowerofLSTMsmay leadtobetterresults.

2017-11-27 IJCNLP 2017 Tutorial 40

Page 41: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

MoreRNNs

• BidirectionalRNN • Stack BidirectionalRNN

2017-11-27 IJCNLP 2017 Tutorial 41

Page 42: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Tree-LSTMs

• TraditionalSequentialComposition

• Tree-StructuredComposition

2017-11-27 IJCNLP 2017 Tutorial 42

Page 43: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

MoreApplicationsofRNN

• NeuralMachineTranslation• HandwritingGeneration• ImageCaptionGeneration• …...

2017-11-27 IJCNLP 2017 Tutorial 43

Page 44: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

ConvolutionNeuralNetwork

CS231nConvolutionalNeuralNetworkforVisualRecognition.

Pooling

2017-11-27 IJCNLP 2017 Tutorial 44

Page 45: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

CNN for NLP

Zhang,Y.,&Wallace,B.(2015).ASensitivityAnalysisof(andPractitioners’Guideto)ConvolutionalNeuralNetworksforSentenceClassification.

2017-11-27 IJCNLP 2017 Tutorial 45

Page 46: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

RecursiveNeuralNetwork

Socher,R.,Manning,C.,&Ng,A.(2011).LearningContinuousPhraseRepresentationsandSyntacticParsingwithRecursiveNeuralNetwork.NIPS.

2017-11-27 IJCNLP 2017 Tutorial 46

Page 47: Part2:Deep Learning - HITir.hit.edu.cn/~car/talks/ijcnlp17-tutorial/lec02-Deep_Learning.pdf · Recurrent Neural Networks (RNNs) •Condition the neural network on all previous inputs

Summary

• Deep Learning– Representation Learning– End-to-end Learning

• Popular Networks– Feedforward Neural Networks– Recurrent Neural Networks– Convolutional Neural Networks

2017-11-27 IJCNLP 2017 Tutorial 47