supervised learning. teacher response: emulation. error: y1 – y2, where y1 is teacher response (...

69
Supervised Learning

Upload: samuel-evans

Post on 21-Jan-2016

227 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Supervised Learning

Page 2: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Teacher response: Emulation.

Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ).Aim: To reduce the error.

It’s Closed Feedback System.

Suppose the error, Error, is on the surface, due to teacher response, we have to bring it down, which is called minimising error, Point is called Local Minimum, or Global Minimum.

Page 3: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Reinforcement Learning.

Page 4: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

There is no teacher.

Converts Primary reinforcement to heuristic reinforcement.

There can be delay for primary reinforcement, as the inputs have to be analysed, which is called credit assignement problem.

Example Character Recognition.

Page 5: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Unsupervised Learning

Page 6: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Error Correction Learning.

Desired Output – Actual Output

Page 7: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Consider Multi Layer Network.

Page 8: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Multi Layer Network, showing output.

Page 9: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Cost Function or Index of Performance.

Widrow Hoff Rule ( Delta Rule )

New Weights

New Weights ( after Unit Delay )

z (power ) – 1, is called unit delay operator.n is disrete time.

Page 10: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Memory Based Learning

Past experinces are stores, to find the relatiion between Input and desired output.

Consider,

Consider

Page 11: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

K Nearest Neighbour

Page 12: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Hebb’s Associative Rules.

Page 13: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Hebb’s Synapse

Page 14: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Four Characteristics of Hebbian Synapse.

Page 15: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 16: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Hebb’s Model

Hebb’s Hypothesis:

Covariance Hypothesis

Increase in inputs, presynapsis, increases outputs ( postsynapsis), leads to saturation.( Activity Product Rule)

Here thresholds are used on inputs and outputs.

Page 17: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 18: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 19: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 20: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 21: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Output Function.

Summation of Weights.

Change in Weights.

Page 22: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Xj is input and xk is output

T is pseudotempearature.

There are two types of neurons, visible and hiddeen

Page 23: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 24: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 25: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 26: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

This is applicable in Error Correction Learning.

Page 27: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

1)Pattern association problemPattern Recognition tasks by Feed Forward Networks

Page 28: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

1) Here every input ( training data ) is associated with an output.

2) So if an input ( test data ) , is close to any training data, like,

Then , Is associated with

3)But if the test data, is very far away from , training data, then Test data, will be associated with an output,

Note: Is very small

And not

4) System displays Accretive Behaviour.

5) Follows Feed Forward Network.

Page 29: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Ai=al + i1, i1 is small number.

Page 30: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

2)Pattern classification problem

Page 31: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

1)In Pattern Association problem, if a set of inputs map to an output, the size of output data set is smaller than input data set. Classes of inputs get a label.2)If a test data, which is close to any inputs ( training data ), in a class, it gets classified , to that class, for which there is a label.3) Here, test data is not associated , with output, but the class has a label, and test data is part of it.4) It creates Accretive behaviour.5) Follows Feed Forward Network.

Page 32: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 33: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

3)Pattern mapping

Page 34: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

1) Here output is a map of input.

2) So if any input ( test data ) , is close to any one training data, the output of test data, will be interpolation of output of training data, means they are in one range.3) Pattern Association and Pattern Classification are derived from Pattern Mapping. Show it by Interpolation.4) Pattern Mapping performs Generalization.5) Follows Feed Forward Network.

Page 35: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Auto Association Problem

Pattern Storage Problem

Pattern Environment Storage Problem

Pattern Recognition tasks by Feed Backward Networks

Page 36: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

1)Auto Association Networks

1) Inputs and Outputs are identical.2) Implementation has to be done by feed backward networks.3)Follows Feed Back Network.

Page 37: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

2)Pattern Storage problem.

The input and output are once again identical.

Three separate neurons are used to realize, the output.So output points and input points are different.Follows Feed Back Network.

Page 38: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

3)Pattern Environment Storage problem

If set of patterns, have certain probability, it is called as pattern environment Storage problem. Follows Feed Back Network.There is feedback, as to getOutput, we have to look at flip of states.

Page 39: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Pattern Recognition tasks by Competitive Learning.

If input patterns are replaced by new patterns, so that, the patterns get the output, over other patterns, it is called as temporary storageProblem. Follows CL. Some Input patterns want to reach output,

1)Temporary Storage problem

Page 40: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

2)Pattern Clustering problem

The test data is classified to the output, based on being near to first class. It creates Accretive Behaviour.Follows CL. Somehow test data, wants to enter testing data. A student, wants to enterer cluster of engg students.

Page 41: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Output is interpolative.Follows CL.Test Data, wants to somehow reach output.

Page 42: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

3)Feature Mapping problem

To cluster we need features by competitive Learning.Ex BP Algorithm.

Page 43: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Back Propagation Algorithm

Page 44: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 45: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 46: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 47: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 48: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Error at output neuron j.

Total Error Energy

1

2

Page 49: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Average of all energies ( at different discrete time intervals )

Activation Values.

3

4

Page 50: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Activation Function.

5

Page 51: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Consider

6

7

8

Page 52: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

9

10

Substituing 7 ,8 ,9 ,10 in 6 we get

11

Page 53: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Using Error Correction Rule

And LMS rule, we get

12

13

Using 11 in 13, we get

14

Page 54: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Where , the error gradient is given by,

15

The above , is to show another way of getting error gradient, to be used in 2nd part.

Page 55: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 56: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 57: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Using the expression in 15,

16

17

Total error at output layer18

Page 58: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

19

20

21

Page 59: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

22

23

24

Page 60: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

From 19

25

26

Substituting 20 in 16, we get

Page 61: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Using 14.

Page 62: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 63: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 64: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

BP Algorithm Summary

Page 65: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 66: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Virtues

Page 67: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 68: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce
Page 69: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce

Limitations ( where brain is better)