2806 neural computation introduction lecture 1

45
2806 Neural Computation Introduction Lecture 1 2005 Ari Visa

Upload: chul

Post on 10-Feb-2016

87 views

Category:

Documents


2 download

DESCRIPTION

2806 Neural Computation Introduction Lecture 1. 2005 Ari Visa. Agenda. Some historical notes Biological background What neural networks are? Properties of neural network Compositions of neural network Relation to artificial intelligence . Overview . - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: 2806 Neural Computation Introduction Lecture 1

2806 Neural ComputationIntroduction

Lecture 1

2005 Ari Visa

Page 2: 2806 Neural Computation Introduction Lecture 1

Agenda

Some historical notes Biological background What neural networks are? Properties of neural network Compositions of neural network Relation to artificial intelligence

Page 3: 2806 Neural Computation Introduction Lecture 1

Overview

The human brain computes in an entirely different way from the conventional digital computer.

The brain routinely accomplishes perceptual recognition in approximately 100-200 ms.

How does a human brain do it?

Page 4: 2806 Neural Computation Introduction Lecture 1

Some Expected Benefits

Nonlinearity Input-Output Mapping Adaptivity Evidential Response Contextual

Information Fault Tolerance

VLSI Implementability

Uniformity of Analysis and Design

Neurobiological Analogy

Page 5: 2806 Neural Computation Introduction Lecture 1

Definition

A neural network is a massive parallel distributed processor made up of simple processing units, which has a natural propensity for storing experimental knowledge and making it available for use. It resembles the brain in two respects:

Page 6: 2806 Neural Computation Introduction Lecture 1

Definition

1) Knowledge is acquired by the network from its environment through a learning process.

2) Interneuron connection strengths, known as synaptic weights, are used to store the acquired knowledge.

Page 7: 2806 Neural Computation Introduction Lecture 1

Some historical notes

Lot of activities concerning automatas, communication, computation, understanding of nervous system during 1930s and 1940s

McCulloch and Pitts 1943von Neumann EDVAC (Electronic Discrete

Variable Automatic Computer)Hebb: The Organization of Behavior, 1949

Page 8: 2806 Neural Computation Introduction Lecture 1

Some historical notes

Page 9: 2806 Neural Computation Introduction Lecture 1

Some historical notes

Minsky: Theory of Neural-Analog Reinforcement Systems and Its Application to the Brain-Model Problem, 1954

Gabor: Nonlinear adaptive filter, 1954 Uttley: leaky integrate and fire neuron,

1956 Rosenblatt: the perceptron, 1958

Page 10: 2806 Neural Computation Introduction Lecture 1

Biological Background

The human nervous system may be viewed as a three stage system (Arbib 1987): The brain continually receives information, perceives it, and makes appropriate decisions.

Page 11: 2806 Neural Computation Introduction Lecture 1

Biological Background

Axons = the transmission lines Dendrites = the receptive zones Action potentials, spikes originate at the

cell body of neurons and then propagate across the individual neurons at constant velocity and amplitude.

Page 12: 2806 Neural Computation Introduction Lecture 1

Biological Background

Synapses are elementary structural and functional units that mediate the interactions between neurons.

Excitation or inhibition

Page 13: 2806 Neural Computation Introduction Lecture 1

Biological Background

Note, that the structural levels of organization are a unique characteristic of the brain

Page 14: 2806 Neural Computation Introduction Lecture 1

Biological Background

Page 15: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

A model of a neuron: synapses (=connecting

links) adder (=a linear

combiner) an activation function

Page 16: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Another formulation of a neuron model

Page 17: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Types of Activation Function:

Threshold Function Piecewise-Linear

Function Sigmoid Function

(signum fuction or hyperbolic tangent function)

Page 18: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Stochastic Model of a NeuronThe activation function of the McCulloch-Pitts

model is given a probabilistic interpretation, a neuron is permitted to reside in only one of two states: +1 or –1. The decision for a neuron to fire is probabilistic.

A standard choice for P(v) is the sigmoid-shaped function = 1/(1+exp(-v/T)), where T is a pseudotemperature.

Page 19: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

The model of an artificial neuron may also be represented as a signal-flow graph.

A signal-flow graph is a network of directed links that are interconnected at certain points called nodes. A typical node j has an associated node signal xj. A typical directed link originates at node j and terminates on node k. It has an associated transfer function (transmittance) that specifies the manner in which the signal yk at node k depends on the signal xj at node j.

Page 20: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Rule 1: A signal flows along a link in the direction defined by the arrow

Synaptic links (a linear input-output relation, 1.19a)

Activation links (a nonlinear input-output relation, 1.19b)

Page 21: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Rule 2: A node signal equals the algebraic sum of all signals entering the pertinent node via the incoming links (1.19c)

Page 22: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Rule 3: The signal at a node is transmitted to each outgoing link originating from that node, with the transmission being entirely independent of the transfer functions of the outgoing links, synaptic divergence or fan-out (1.9d)

Page 23: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network A neural network is a directed

graph consisting of nodes with interconnecting synaptic and activation links, and is characterized by four properties:

1. Each neuron is represented by a set of linear synaptic links, an externally applied bias, and a possibly nonlinear activation link, This bias is represented by a synaptic link connected to an input fixed at +1.

Page 24: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

2. The synaptic links of a neuron weight their respective input signals.

3. The weighted sum of the input signals defines the induced local field of a neuron in question.

4. The activation link squashes the induced local field of the neuron to produce an output.

Page 25: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Complete graph Partially complete

graph = architectural graph

Page 26: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

Feedback is said to exist in a dynamic system whenever the output of an element in the system influences in part the input applied to the particular element, thereby giving rise to one or more closed paths for the transmission of signals around the system (1.12)

Page 27: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

yk(n) = A[x’j(n)] x’j(n) = xj(n)+B[yk(n)] yk(n)=A/(1-AB)[xj(n)] the closed-loop

operator A/(1-AB) the open-loop operatorABIn general AB BA

Page 28: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network

A/(1-AB) w/1-wz-1)

yk(n) is convergent (=stable), if |w| < 1 (1.14a)

yk(n) is divergent (=unstable), if |w| < 1

0

1 )1()(l

jl

k nxwny

Page 29: 2806 Neural Computation Introduction Lecture 1

Properties of Neural Network A/(1-AB) w/1-wz-1)

yk(n) is convergent (=stable), if |w| < 1 (1.14a)

yk(n) is divergent (=unstable), if |w| 1,

if |w| = 1 the divergence is linear (1.14.b)

if |w| >1 the divergence is exponential (1.14c)

0

1 )1()(l

jl

k nxwny

Page 30: 2806 Neural Computation Introduction Lecture 1

Compositions of Neural Network

The manner in which the neurons of a neural network are structured is intimately linked with the learning algorithm used to train the network.

Single-Layer Feedforward Networks

Page 31: 2806 Neural Computation Introduction Lecture 1

Compositions of Neural Network

Multilayer Feedforward Networks (1.16)

Hidden layers, hidden neurons or hidden units -> enabled to extract higher-order statistics

Page 32: 2806 Neural Computation Introduction Lecture 1

Compositions of Neural Network

Recurrent Neural Network (1.17)

It has at least one feedback loop.

Page 33: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

Knowledge refers to stored information or models used by a person or machine to interpret, predict, and appropriately respond to the outside world (Fishler and Firschein, 1987)

A major task for neural network is to learn a model of the world

Page 34: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

Knowledge of the world consists of two kind of information

1) The known world state, prior information 2) Observations of the world, obtained by

means of sensor. Obtained observations provide a pool of

information from which the examples used to train the neural network are drawn.

Page 35: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

The examples can be labelled or unlabelled. In labelled examples, each example representing

an input signal is paired with a corresponding desired response. Note, both positive and negative examples are possible.

A set of input-output pairs, with each pair consisting of an input signal and the corresponding desired response, is referred to as a set of training data or training sample.

Page 36: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

Selection of an appropriate architecture A subset of examples is used to train the

network by means of a suitable algorithm (=learning).

The performance of the trained network is tested with data not seen before (=testing).

Generalization

Page 37: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

Rule 1: Similar inputs from similar classes should usually produce similar representations inside the network, and should therefore be classified as belonging to the same category.

Rule 2: Items to be categorized as separate classes should be given widely different representations in the network.

Page 38: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

Rule 3: If a particular feature is important, then there should be a large number of neurons involved in the representation of that item in the network

Rule 4: Prior information and invariances should be built into the design of a neural network, thereby simplifying the network design by not having to learn them.

Page 39: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

How to Build Prior Information Into Neural Network Design?

1) Restricting the network architecture through the use of local connections known as receptive fields.

2) Constraining the choice of synaptic weights through the use of weight-sharing.

Page 40: 2806 Neural Computation Introduction Lecture 1

Knowledge Representation

How to Build Invariances into Neural Network Design?

1) Invariance by Structure 2) Invariance by Training 3) Invariant Feature Space

Page 41: 2806 Neural Computation Introduction Lecture 1

Relation to Artificial Intelligence

The goal of artificial intelligence (AI) is the development of paradigms or algorithms that require machines to perform cognitive tasks (Sage 1990).

An AI system must be capable of doing three things:

1) Store knowledge 2) Apply the knowledge stored to solve problems. 3) Acquire new knowledge through experience.

Page 42: 2806 Neural Computation Introduction Lecture 1

Relation to Artificial Intelligence

Representation: The use of a language of symbol structures to represent both general knowledge about a problem domain of interest and specific knowledge about the solution to the problem

Declarative knowledgeProcedural knowledge

Page 43: 2806 Neural Computation Introduction Lecture 1

Reasoning: The ability to solve problems The system must be able to express and solve a broad range of problems and

problem types. The system must be able to make explicit and implicit information known to it. The system must have control mechanism that determines which operations to

apply to a particular problem.

Relation to Artificial Intelligence

Page 44: 2806 Neural Computation Introduction Lecture 1

Relation to Artificial Intelligence Learning: The

environment supplies some information to a learning element. The learning element then uses this information to make improvements in a knowledge base, and finally the performance element uses the knowledge base to perform its task.

Page 45: 2806 Neural Computation Introduction Lecture 1

Summary

A major task for neural network is to learn a model of the world

It is not a totally new approach but it has differences to AI, matematical modeling, Pattern Recognition and so on.