chapter 1: intelligent agents · francine (penny) patterson (1947 { ) ronald cohn (?) koko (1971 {...

94
Chapter 1: Intelligent Agents c D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 1

Upload: others

Post on 13-Aug-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Chapter 1: Intelligent Agents

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 1

Page 2: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Lecture Overview

What is Artificial Intelligence?

Agents acting in an environment

Learning objectives: at the end of the class, you should beable to

describe what an intelligent agent is

identify the goals of Artificial Intelligence

classify the inputs and the outputs of various agents

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 2

Page 3: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

What is Artificial Intelligence?

Artificial Intelligence is the synthesis and analysis ofcomputational agents that act intelligently.

An agent is something that acts in an environment.

An agent acts intelligently if:I its actions are appropriate for its goals and circumstancesI it is flexible to changing environments and goalsI it learns from experienceI it makes appropriate choices given perceptual and

computational limitations

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 3

Page 4: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Observing and investigating intelligence

reflexes

insticts

problem solving

learning

tool use

intellectual tasks

language communication

creativity

self-recognition

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 4

Page 5: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Reflexes

involuntary and nearly instantaneous movement in response toa stimulus

can be innate or aquired (conditioning)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 5

Page 6: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Reflexes

Can intelligence emerge from the interplay of reflexes?

Is building structure always a sign of intelligent behaviour?

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 6

Page 7: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Instincs

Migrating animals: birds, butterflies, mammals, ...

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 7

Page 8: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Instincs

the inherent inclination of a living organism toward a particularcomplex behavior.

innate behavior: absence of learning

eating, nest building, mating, feeding, migration, ...

no intelligent behaviour, can be counterproductive

can be overridden by competing instincts or reasoning

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 8

Page 9: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Problem solving

Wolfgang Kohler (1887 – 1967)

Intelligenzprufungen an Anthropoiden (1917)The Mentality of Apes (1925)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 9

Page 10: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Problem solving

behaviour is driven by insight, if a goal cannot be reachedon a direct and simple path, but a way round is availableand chosen.

distinction between true intelligent results andcoincidental success.

I chance: several independent movementsI true intelligence: spatially and temporally coherent

course without hesitation

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 10

Page 11: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Problem solving

Robert Epstein (1953 – )

are pigeons as intelligent as chimpanzees?

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 11

Page 12: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Problem solving

Bernd Heinrich (1953 – )

problem solving in ravens

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 12

Page 13: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Problem solving

Making the task more challenging ...

Animals mastered the second task only after they have solvedthe first one!

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 13

Page 14: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Conditioning

Iwan Petrowitsch Pawlow (1849 – 1936)

acquired/learned reflexes

association learning: direct linking ofsensual stimuli

assumed to be the only type of learning in animals

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 14

Page 15: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Imprinting

phase-sensitive learning

rapid

apparently independent of the consequences of thebehavior

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 15

Page 16: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Imitation

observing and replicating another individuums behavior

form of social learning

leads to the development of traditions and culture

transfer of information (behaviours, customs, etc.) acrossgenerations without the need for genetic inheritance

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 16

Page 17: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Imitation

blue tits learned to open sealed milk bottles

1900 – 1945 open milk bottles were dropped at the doorsteparound 1945 sealed milk bottles have been introducedsince 1960 doorstep delivery has been suspended

around 1945 first indiviuals learned to open the sealaround 1950 the whole population of

British blue tits had aquiredthe skill

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 17

Page 18: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Imitation

Are blue tits smarter than european robins?

Only few individuals acquired the skill but not the wholepopulation. Why?

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 18

Page 19: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Imitation

Are blue tits smarter than ravens?

The string pulling experiment again:

Every individual developed its own solution strategy

No transfer of skills between individuals

One raven didn’t master the task even though it was ableto observe the other ones pulling the meat

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 19

Page 20: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Tool use

Jane Goodall (1934 – )

Tool-Using and Aimed Throwing in aCommunity of Free-Living Chimpanzees (1964)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 20

Page 21: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intellectual tasks

Der kluge Hans (1895 – 1916 ?)

Orlow-Traber

owner: Wilhelm von Osten (1838 – 1909)

ability to count and calculateand to distinguish simple concepts

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 21

Page 22: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intellectual tasks

Has Hans been intelligent?

Hans commission (1904): no manipulation found

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 22

Page 23: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intellectual tasks

Oskar Pfungst (1874 – 1933)

Rigorous testing:

Isolating horse and questionerfrom spectators

Using questioners other thanthe horse’s master

Blindfolding the horse

Using questions, where the questioner didn’t know theanswer

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 23

Page 24: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intellectual tasks

Hans could get the correct answer even if not von Ostenwas asking

horse must have seen the questioner

questioner had to know what the answer was

von Osten knew the answer → 89% correct

von Osten did not know the answer → only 6% correct

the questioner’s posture and facial expression changed asthe horse’s taps approached the right answer

the changes were consistent with an increase in tension

the tension was released when the horse made the final,correct tap

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 24

Page 25: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intellectual tasks

Confirming the findings by inverse experiments:

cues from the audience are sufficiently reliable

⇒ the Clever Hans effect

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 25

Page 26: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intellectual tasks

Irene Pepperberg (1949 –)Teaching numerical concepts to parrots

counting (up to seven)

simple inferences on numbers

cardinal vs. ordinal numbers

the notion of zero

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 26

Page 27: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Natural language communication

Alex the grey parrot

Washoe the chimp

Koko the gorilla

Nim Chimpsky

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 27

Page 28: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Alex the grey parrot

Irene Pepperberg (1949 –)Alex (1976 – 2007)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 28

Page 29: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Alex the grey parrot

MAN Come on, what is it?ALEX Keychain.

I.P. Good birdie. Good parrot. What is it?ALEX Rock.MAN Good boy.

I.P. Yeah, good birdie. Alex, what toy?ALEX Nail.

I.P. Nail, that’s right. You’re a good birdie. You’re a very good boy.MAN What toy?ALEX Truck.

I.P. That’s right.MAN You’re a very good birdie.

I.P. Tell me what color. What color?ALEX Yellow.

I.P. Yellow, that’s right.MAN What matter?ALEX Wood.MAN Good. That’s right. Very good.

I.P. How many? Good boy. How many?ALEX Two.

I.P. Good parrot. Good boy. One. Two.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 29

Page 30: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Alex the grey parrot

I.P. Can you tell me what’s different? What’s different?ALEX Color.

I.P. Good boy. All right. What same? What same?ALEX Shape.

I.P. Good boy, good birdie. What color bigger? You know. What colorbigger?

ALEX Yellow.I.P. Good boy. Good birdie.I.P. Look. What matter four-corner blue?

DENISE What matter four-corner blue?ALEX Wood.

DENISE That’s a good boy. You’re right.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 30

Page 31: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Alex the grey parrot

After 19 years of training:

200/500 lexical items (active/passive)

basic language understanding and language productioncapabilities

different dimensions of object descriptions (color, shape,material, ...)

complex object descriptions: four-corner wood

refusal to cooperate; rejection of food or toys

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 31

Page 32: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Washoe the chimp

Allen Gardner (1930 –)Beatrice Gardner (1934 – 1995)Washoe (1965 – 2007)

Teaching American Sign Language

raised like a child

private trailer with living and cooking areas

learned 350 words of ASL

taught her adopted son Loulis some ASL

invention of new sign combinations: swan = water + bird

simple verb-noun combinations

no conditioning, no rewards

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 32

Page 33: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Koko the gorilla

Francine (Penny) Patterson (1947 – )Ronald Cohn (?)Koko (1971 – )

production: 1000 signs, comprehension: 2000 signs

learning or conditioning?

invention of new sign combinations: ring = finger +bracelet

no sentences, but adjectives, nouns and noun phrases

rewards if answer is somehow ok: apple is apple or red

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 33

Page 34: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Nim Chimpsky

Herbert S. Terrace

Nim Chimpsky (1973 – 2000)

not ”raised like a child”

learned 150 signs

but Terrace concluded that thiswas not natural language

I mean length of 20000 recordedresponses only 1.2

I no correlation between lexical growth and structuralcomplexity

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 34

Page 35: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Nim Chimpsky

strong criticism of the experimental methodology

apes remained passive

interpretation of responses is up to the experimenter

human language capabilities are projected onto the ape

often the response of the ape was preceded by an(unvoluntary) similar movement of the experimenter (250ms earlier)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 35

Page 36: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Creativity

Music playing robots

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 36

Page 37: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Creativity

Painting elephants

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 37

Page 38: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Creativity

Desmond Morris (1928 –)Painting chimpanzees

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 38

Page 39: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Self-recognition

Does the animal recognize itself as an acting agent?⇒ mirror test

Passing the test:

Humans (at age 18 months or older), Bonobos, Chimpanzees,Orangutans, Gorillas (Koko!), Bottlenose dolphins, Orcas,Elephants, European Magpies (the only non-mammal)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 39

Page 40: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Conclusions

intelligence is a composite propertyI different kinds of intelligence

intelligence is a graded notionI cognitive tasks with different degrees of difficulty

intelligence is a highly subjective notionI contradictory interpretations of observations are quite

common

danger of ”anthropo”morphizationI objective investigation methods required

studying intelligence requires systematic analysisI no conclusions can be drawn from isolated observations

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 40

Page 41: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Goals of Artificial Intelligence

Scientific goal: to understand the principles that makeintelligent behavior possible in natural or artificialsystems.

I analyze natural and artificial agentsI formulate and test hypotheses about what it takes to

construct intelligent agentsI design, build, and experiment with computational

systems that perform tasks that require intelligence

Engineering goal: design useful, intelligent artifacts.

Analogy between studying flying machines and thinkingmachines.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 41

Page 42: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Agents acting in an environment

Environment

ObservationsActions

Past Experiences

Goals/Preferences

Prior Knowledge Agent

Abilities

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 42

Page 43: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Examples of Agents

Organisations Microsoft, Al Qaeda, Government ofCanada, UBC, CS Dept,...

People teachers, physicians, stock traders, engineers,researchers, travel agents, farmers, waiters...

Computers/devices thermostats, user interfaces, airplanecontrollers, network controllers, games, advising systems,tutoring systems, diagnostic assistants, robots, Googlecar, Mars rover...

Animals dogs, mice, birds, insects, worms, bacteria...

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 43

Page 44: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Inputs to an agent

Abilities — the set of things it can do

Goals/Preferences — what it wants, its desires, itsvalues,...

Prior Knowledge — what it comes into being knowing,what it doesn’t get from experience,...

History of observations (percepts, stimuli) of theenvironment

I (current) observations — what it observes nowI past experiences — what it has observed in the past

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 44

Page 45: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Example agent: robot

abilities: movement, grippers, speech, facialexpressions,. . .

goals: deliver food, rescue people, score goals,explore,. . .

prior knowledge: what is important feature, categories ofobjects, what a sensor tell us,. . .

observations: vision, sonar, sound, speech recognition,gesture recognition,. . .

past experiences: effect of steering, slipperiness, howpeople move,. . .

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 45

Page 46: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Example agent: teacher

abilities: present new concept, drill, give test, explainconcept,. . .

goals: particular knowledge, skills, inquisitiveness, socialskills,. . .

prior knowledge: subject material, teaching strategies,. . .

observations: test results, facial expressions, errors,focus,. . .

past experiences: prior test results, effects of teachingstrategies, . . .

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 46

Page 47: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Agents acting in an environment

Environment

ObservationsActions

Past Experiences

Goals/Preferences

Prior Knowledge Agent

Abilities

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 47

Page 48: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Dimensions

Research proceeds by making simplifying assumptions,and gradually reducing them.

Each simplifying assumption gives a dimension ofcomplexity

I multiple values in a dimension: from simple to complexI simplifying assumptions can be relaxed in various

combinations

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 48

Page 49: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Dimensions of Complexity

Flat or modular or hierarchical

Explicit states or features or individuals and relations

Static or finite stage or indefinite stage or infinite stage

Fully observable or partially observable

Deterministic or stochastic dynamics

Goals or complex preferences

Single-agent or multiple agents

Knowledge is given or knowledge is learned fromexperience

Perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 49

Page 50: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Modularity

Model at one level of abstraction: flat

Model with interacting modules that can be understoodseparately: modular

Model with modules that are (recursively) decomposedinto modules: hierarchical

Example: Planning a trip from here to a see the MonaLisa in Paris.

Flat representations are adequate for simple systems.

Complex biological systems, computer systems,organizations are all hierarchical

A flat description is either continuous or discrete.Hierarchical reasoning is often a hybrid of continuous anddiscrete.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 50

Page 51: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Succinctness and Expressiveness

Much of modern AI is about finding compact representationsand exploiting the compactness for computational gains.A agent can reason in terms of:

Explicit states — a state is one way the world could be

Features or propositions.I States can be described using features.I 30 binary features can represent 230 = 1, 073, 741, 824

states.

Individuals and relationsI There is a feature for each relationship on each tuple of

individuals.I Often an agent can reason without knowing the

individuals or when there are infinitely many individuals.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 51

Page 52: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Planning horizon

...how far the agent looks into the future when deciding whatto do.

Static: world does not change

Finite stage: agent reasons about a fixed finite numberof time steps

Indefinite stage: agent reasons about a finite, but notpredetermined, number of time steps

Infinite stage: the agent plans for going on forever(process oriented)

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 52

Page 53: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Uncertainty

There are two dimensions for uncertainty. In each dimensionan agent can have

No uncertainty: the agent knows which world is true

Disjunctive uncertainty: there is a set of worlds that arepossible

Probabilistic uncertainty: a probability distribution overthe worlds.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 53

Page 54: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Why Probability?

Agents need to act even if they are uncertain.

Predictions are needed to decide what to do:I definitive predictions: you will be run over tomorrowI disjunctions: be careful or you will be run overI point probabilities: probability you will be run over

tomorrow is 0.002 if you are careful and 0.05 if you arenot careful

I probability ranges: you will be run over with probabilityin range [0.001,0.34]

Acting is gambling: agents who don’t use probabilitieswill lose to those who do.

Probabilities can be learned from data and priorknowledge.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 54

Page 55: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Uncertain dynamics

If an agent knew the initial state and its action, could itpredict the resulting state?The dynamics can be:

Deterministic : the resulting state is determined from theaction and the state

Stochastic : there is uncertainty about the resultingstate.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 55

Page 56: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Sensing Uncertainty

Whether an agent can determine the state from itsobservations:

Fully-observable : the agent can observe the state of theworld.

Partially-observable : there can be a number states thatare possible given the agent’s observations.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 56

Page 57: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Goals or complex preferences

achievement goal is a goal to achieve. This can be acomplex logical formula.

complex preferences may involve tradeoffs betweenvarious desiderata, perhaps at different times.

I ordinal only the order mattersI cardinal absolute values also matter

Examples: coffee delivery robot, medical doctor

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 57

Page 58: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Single agent or multiple agents

Single agent reasoning is where an agent assumes thatany other agents are part of the environment.

Multiple agent reasoning is when an agent reasonsstrategically about the reasoning of other agents.

Agents can have their own goals: cooperative, competitive, orgoals can be independent of each other

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 58

Page 59: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Learning from experience

Whether the model is fully specified a priori:

Knowledge is given.

Knowledge is learned from data or past experience.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 59

Page 60: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Perfect rationality or bounded rationality

Perfect rationality: the agent can determine the bestcourse of action, without taking into account its limitedcomputational resources.

Bounded rationality: the agent must make gooddecisions based on its perceptual, computational andmemory limitations.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 60

Page 61: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Dimensions of Complexity

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single-agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 61

Page 62: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

State-space Search

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 62

Page 63: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Classical Planning

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 63

Page 64: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Decision Networks

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 64

Page 65: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Markov Decision Processes (MDPs)

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 65

Page 66: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Decision-theoretic Planning

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 66

Page 67: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Reinforcement Learning

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 67

Page 68: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Classical Game Theory

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 68

Page 69: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Humans

flat or modular or hierarchical

explicit states or features or individuals and relations

static or finite stage or indefinite stage or infinite stage

fully observable or partially observable

deterministic or stochastic dynamics

goals or complex preferences

single agent or multiple agents

knowledge is given or knowledge is learned

perfect rationality or bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 69

Page 70: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

The Dimensions Interact in Complex Ways

Partial observability makes multi-agent and indefinitehorizon reasoning more complex

Modularity interacts with uncertainty and succinctness:some levels may be fully observable, some may bepartially observable

Three values of dimensions promise to make reasoningsimpler for the agent:

I Hierarchical reasoningI Individuals and relationsI Bounded rationality

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 70

Page 71: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Four Example Application Domains

Autonomous delivery robot roams around an officeenvironment and delivers coffee, parcels,. . .

Diagnostic assistant helps a human troubleshootproblems and suggests repairs or treatments. E.g.,electrical problems, medical diagnosis.

Intelligent tutoring system teaches students in somesubject area.

Trading agent buys goods and services on your behalf.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 71

Page 72: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Domain for Delivery Robot

lab4

stairs

lab1 lab2

lab3

r101 r103 r105 r107 r109 r111

r113

r115

r117

r119r121r123r125r127r129r131

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 72

Page 73: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Autonomous Delivery Robot

Example inputs:

Abilities: movement, speech, pickup and place objects.

Prior knowledge: its capabilities, objects it mayencounter, maps.

Past experience: which actions are useful and when,what objects are there, how its actions affect its position.

Goals: what it needs to deliver and when, tradeoffsbetween acting quickly and acting safely.

Observations: about its environment from cameras,sonar, sound, laser range finders, or keyboards.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 73

Page 74: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

What does the Delivery Robot need to do?

Determine where Craig’s office is. Where coffee is. . .

Find a path between locations.

Plan how to carry out multiple tasks.

Make default assumptions about where Craig is.

Make tradeoffs under uncertainty: should it go near thestairs?

Learn from experience.

Sense the world, avoid obstacles, pickup and put downcoffee.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 74

Page 75: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Domain for Diagnostic Assistant

light

two-wayswitch

switchoff

on

poweroutlet

circuit breaker

outside powercb1

s1

w1s2 w2

w0

l1

w3s3

w4

l2p1

w5

cb2

w6

p2

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 75

Page 76: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Diagnostic Assistant

Example inputs:

Abilities: recommends fixes, ask questions.

Prior knowledge: how switches and lights work, howmalfunctions manifest themselves, what information testsprovide, the side effects of repairs.

Past experience: the effects of repairs or treatments, theprevalence of faults or diseases.

Goals: fixing the device and tradeoffs between fixing orreplacing different components.

Observations: symptoms of a device or patient.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 76

Page 77: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Subtasks for the diagnostic assistant

Derive the effects of faults and interventions.

Search through the space of possible fault complexes.

Explain its reasoning to the human who is using it.

Derive possible causes for symptoms; rule out othercauses.

Plan courses of tests and treatments to address theproblems.

Reason about the uncertainties/ambiguities givensymptoms.

Trade off alternate courses of action.

Learn what symptoms are associated with faults, theeffects of treatments, and the accuracy of tests.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 77

Page 78: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Trading Agent

Trading agent interacts with an information environment topurchase goods and services.

It acquires a users needs, desires and preferences.

It finds what is available.

It purchases goods and services that fit together to fulfillyour preferences.

It is difficult because users preferences and what isavailable can change dynamically, and some items may beuseless without other items.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 78

Page 79: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Trading Agent Inputs

Abilities: acquire information, make recommendations,purchase items.

Prior knowledge: the ontology of what things areavailable, where to purchase items, how to decompose acomplex item.

Past experience: how long special last, how long itemstake to sell out, who has good deals, what yourcompetitors do.

Goals: what the person wants, their tradeoff.

Observations: what items are available, prices, numberin stock,

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 79

Page 80: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Intelligent Tutoring System

Abilities: Present information, give tests

Prior knowledge: subject material, primitive strategies

Past experience: common errors, effects of teachingstrategies

Goals: the students should master subject material, gainsocial skills, study skills, inquisitiveness, interest

Observations: test results, facial expressions, questions,what the student is concentrating on

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 80

Page 81: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Common Tasks of the Domains

Modeling the environment Build models of the physicalenvironment, patient, or information environment.

Evidential reasoning or perception Given observations,determine what the world is like.

Action Given a model of the world and a goal, determinewhat should be done.

Learning from past experiences Learn about the specificcase and the population of cases.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 81

Page 82: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Representations

problem

representation

solution

output

solve

compute

informal

formalrepresent interpret

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 82

Page 83: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

What do we want in a representation?

We want a representation to be

rich enough to express the knowledge needed to solve theproblem;

as close to the problem as possible: compact, natural andmaintainable;

amenable to efficient computationI able to express features of the problem that can be

exploited for computational gainI able to trade off accuracy and computation time and/or

space

able to be acquired from people, data and pastexperiences.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 83

Page 84: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Defining a Solution

Given an informal description of a problem, what is asolution?

Typically much is left unspecified, but the unspecifiedparts can’t be filled in arbitrarily.

Much work in AI is motivated bycommon-sense reasoning. The computer needs to make

common-sense conclusions about the unstatedassumptions.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 84

Page 85: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Quality of Solutions

Does it matter if the answer is wrong or answers aremissing?

Classes of solution:

An optimal solution is a best solution according somemeasure of solution quality.

A satisficing solution is one that is good enough,according to some description of which solutions areadequate.

An approximately optimal solution is one whose measureof quality is close to the best theoretically possible.

A probable solution one that is likely to be a solution.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 85

Page 86: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Decisions and Outcomes

Good decisions can have bad outcomes. Bad decisionscan have good outcomes.

Information can be valuable because it leads to betterdecisions: value of information.

We can often trade off computation time and solutionquality.An anytime algorithm can provide a solution at anytime; given more time it can produce better solutions.

An agent isn’t just concerned about finding the right answer,but about acquiring the appropriate information, andcomputing it in a timely manner.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 86

Page 87: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Solution quality and computation time

0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5

discounted quality

absolute

time discount

time

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 87

Page 88: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Choosing a Representation Language

We need to represent a problem to solve it on a computer. problem→ specification of problem→ appropriate computation

Example representation languages: Machine Language, C++,Java, Prolog, English

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 88

Page 89: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Hierarchy of representations

problem solution

high-level specification

programming language

assembly language

digital circuit

currents and voltages

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 89

Page 90: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Physical symbol system hypothesis

A symbol is a meaningful physical pattern that can bemanipulated.

A symbol system creates, copies, modifies and destroyssymbols.

Physical symbol system hypothesis:

A physical symbol system has the necessary and sufficientmeans for general intelligent action.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 90

Page 91: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Knowledge & Symbol Levels

Two levels of abstraction seem to be common amongbiological and computational entities:

The knowledge level is in terms of what an agent knowsand what its goals are.

The symbol level is a level of description of an agent interms of what reasoning it is doing.

The knowledge level is about the external world to the agent.The symbol level is about what symbols an agent uses toimplement the knowledge level.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 91

Page 92: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Mapping from Problem to Representation

What level of abstraction of the problem to represent?What individuals and relations in the world to represent?How can an agent represent the knowledge to ensure thatthe representation is natural, modular, and maintainable?How can an agent acquire the information from data,sensing, experience, or other agents?

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 92

Page 93: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Choosing a level of abstraction

A high-level description is easier for a human to specifyand understand.

A low-level description can be more accurate and morepredictive. High-level descriptions abstract away detailsthat may be important for actually solving the problem.

The lower the level, the more difficult it is to reason with.

You may not know the information needed for a low-leveldescription.

It is sometime possible to use multiple levels of abstraction.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 93

Page 94: Chapter 1: Intelligent Agents · Francine (Penny) Patterson (1947 { ) Ronald Cohn (?) Koko (1971 { ) production: 1000 signs, comprehension: 2000 signs learning or conditioning? invention

Reasoning and acting

Reasoning is the computation required to determine what anagent should do.

Design time reasoning and computation is carried out bythe designer the agent.

Offline computation is the computation done by theagent before it has to act.Background knowledge and data knowledge base .

Online computation is the computation that’s done byan agent between receiving information and acting.

c©D. Poole, A. Mackworth 2010, W. Menzel 2013 Artificial Intelligence, Lecture 1.1, Page 94