ie 585 associative network. 2 associative memory nn single-layer net in which the weights are...

21
IE 585 Associative Network

Upload: eric-hamilton

Post on 02-Jan-2016

212 views

Category:

Documents


0 download

TRANSCRIPT

IE 585

Associative Network

2

Associative Memory NN

• Single-layer net in which the weights are determined in such a way that the net can store a set of pattern associations

• The net only only learns the specific pattern pairs that were used for training, but also is able to recall the desired response pattern when given an input stimulus that is similar, but not identical, to the training input

3

Types of Associative Memory

• Autoassociative memory:

if each vector t (output, target) is the same as the vector s (input) with which it is associated

• Heteroassociative memory:

if the t’s are different from the s’s.

4

Discrete Hopfield Net

• Developed by Hopfield (binary – 1982, bipolar – 1984)

• Recurrent autoassociative• Fully interconnected neural network• Symmetric weights with no self-connections• Asychronous

only one unit updates its activation at a time each unit continues to receive an external signal in addition to the signal from the other units in the net

5

Architecture of Hopfield Net

x1

x2

Y1

.

.xn

Y2

Yn

6

Procedure of the Discrete Hopfield Net

Initialize weights to store patterns

For each input vector x

set yi = xi

for each unit Yi

compute

broadcast the yi to all other units

Continue until it converges

j

jjiiY ywxneti

7

Hebb Learning• Binary

• Bipolar

1)(21)(2)1( pspsw jp

iijij

p

jiijij pspsw )()()1(

ji

ji

ij

ijij if 0

if 1function Kronecker :

8

Transfer Function

• Step Function

Binary

Transfer Function:

Bipolar

Transfer Function:

1

1

0

-1

iY

iY

iY

ii

i

i

i

net

net

net

yy

if

if

if

0

1

iY

iY

iY

ii

i

i

i

net

net

net

yy

if

if

if

1

1

9

i• is usually 0

• The order of update of the units is random

• The order of learning set does not affect

• Extension to continuous activation for both pattern association or constrained optimization by Hopfiled & Tank (1985, Hopfiled-Tank Net)

10

Hopfield Net Example

11

Energy Function for the discrete Hopfield Net

• Also called Lyapunov Function• Developed by Alexander Lyaphnov (Russian)• Asynchronous updating of the Hopfield net

allows such a function• Prove that the net will converge to a stable

limit point (pattern of activation of the units)• Non-increasing function of the state of the

system

12

Storage Capacity of Hopfield Net

• Binary

• Bipolar

P: # of patterns that can be stored an recalled in a net with reasonable accuracyn: # of neurons in the net

nP 15.0

n

nP

2log2

13

Spurious Memory

• It is stable in energy state, but converges to an activation vector that is not one of the stored patterns

a discrete Hopfield net can be used to determine whether an input vector is a “known” vector or an “unknown” vector

14

• Hamming Distance (HD)

# of different bits in two binary or bipolar vectors

• Orthgonal

HD = n/2 n: # of bits

can store max # of patterns

15

Bi-directional Associative Memory (BAM)

• Developed by Kosko (1988)• Recurrent heteroassociative• Two layers of neurons connected by directional

weighted connection paths• Symmetric weights with no self-connections• Signals are sent only from one layer to the other

at any step of the process, not simultaneously in both directions

16

Architecture of BAM

x1

x2

Y1

.

.

xn

Y2

Yn

.

.

17

Procedure of the BAMInitialize weights to store patterns

Initialize all activations to 0

For each testing input vector

present input x to the X-layer (or Y-layer)

compute

compute

send signals to Y-layer

Continue until it converges

i

iijY xwnetj

)(jYj netfy

j

jijX ywneti

)(iXi netfx

18

Hebb Learning• Binary

• Bipolar

1)(21)(2 ptpsw jp

iij

p

jiij ptpsw )()(

19

Transfer Function

• Step Function

Binary

Transfer Function:

Bipolar

Transfer Function:

1

1

0

-1

jY

jY

jY

jj

j

j

j

net

net

net

yy

if

if

if

0

1

jY

jY

jY

jj

j

j

j

net

net

net

yy

if

if

if

1

1

20

BAM Example

21

Storage Capacity of the BAM

P: # of patterns that can be stored an recalled in a net with reasonable accuracyn: # of inputsm: # of outputs

),min( mnP