ling 388: language and computers sandiway fong lecture 24: 11/17

23
LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Upload: jason-heath

Post on 29-Dec-2015

215 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

LING 388: Language and Computers

Sandiway Fong

Lecture 24: 11/17

Page 2: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Administrivia

• Homework #5– WordNet– out today– due Tuesday 29th (after Thanksgiving)

Page 3: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Last Time

• semantic networks based around language– WordNet (Miller @ Princeton University)

• handbuilt (ad hoc) network of synonym sets (synsets) connected by semantic relations

• e.g. isa, part of, antonymy, causation etc.

• large-scale (and free) lexical resource– 139,000 entries (word senses) v1.7– 10,000 verbs (polysemy 2)– 20,000 adjectives (1.5)

Page 4: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Last Time

• semantic networks based around language– WordNet (Miller @ Princeton University)

• handbuilt (ad hoc) network of synonym sets (synsets) connected by semantic relations

• e.g. isa, part of, antonymy, causation etc.

• example (semantic opposition):– an instance of the frame problem– John mended the torn/red dress

– mend as a change-of-state verb: how to track changes? mend: x CAUS y BECOME <STATE (mended)>

– John CAUS the torn/red dress BECOME <STATE (mended)>

Page 5: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Last Time

• example (semantic opposition):• what kind of knowledge is invoked here?

– exploit the network

– or– some kind of enriched lexicon with encyclopedic knowledge

Page 6: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Last Time

• semantic networks based around language– WordNet (Miller @ Princeton University)

• handbuilt (ad hoc) network of synonym sets (synsets) connected by semantic relations

• e.g. isa, part of, antonymy, causation etc.

• example (logical metonomy):– an instance of filling in the missing verb– John began the novel– John began [reading/writing] the novel– defeasible– my goat eats anything– he really enjoyed [verb] your book (reading) (eating)

Page 7: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Last Time

• example (logical metonomy):– compute some telic role– John enjoyed [drinking] the wine

– Generative Lexicon (GL) Model (wine: telic role: drink) – vs.

– WordNet structure

Page 8: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Larger Issue

• larger issue– knowledge of language for semantic inference– useful for sentence understanding applications

• questions– how do we organize such knowledge?– do we have a semantic network like WordNet?– is it part of the “language faculty”?

Page 9: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Homework 5

Page 10: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Homework 5: Question 1

• (3pts) use WordNet to discover– what have (specific senses) of the following nouns in

common?• Umbrella• Saucepan• Baseball bat• Carpet beater

– but do not share with:• Giraffe• Pretzel• Homework

– explain your answer

Page 11: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

GRE

• Educational Testing Service (ETS)– www.ets.org– 13 million standardized tests/year– Graduate Record Examination (GRE)

• verbal section of the GRE– vocabulary

• GRE vocabulary– word list

• Word list retention– word matching exercise (from a GRE prep book)

Page 12: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Task: Match each word in the first column with its definition in the second column

accolade

abate

aberrant

abscond

acumen

abscission

acerbic

accretion

abjure

abrogate

deviating

abolish

keen insight

lessen in intensity

sour or bitter

building up

depart secretly

renounce

removal

praise

Page 13: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Homework 5: Question 2

• (for 17 pts)• use WordNet (wnconnect) to “solve” the word

match puzzle– come up with an method or procedure that correctly

associates (as many as possible)words in the left column to those on the rightyou can use different forms, e.g. deviate/deviation/deviating

• describe your procedure – show how it works on the examples– (you don’t have to turn in a program)– e.g. you could compute shortest links

Page 14: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Homework 5: Question 2

• wnconnect– compute links

between abate and lessen

abate and sour

Page 15: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Task: match each word in the first column with its definition in the second column

accolade

abate

aberrant

abscond

acumen

abscission

acerbic

accretion

abjure

abrogate

deviation

abolish

keen insight

lessen in intensity

sour or bitter

building up

depart secretly

renounce

removal

praise

Page 16: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Language and Intelligence

• if a computer program can be written to do as well as humans on examples like homework 5, is the program intelligent?

• can such a program be written?

Page 17: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

• a look at the future of educational testing...• www.etstechnologies.com• e-rater

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Page 18: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

• www.ets.org/erater/index.html

Page 19: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

• e-rater FAQs• Q. What's the technology used in e-rater?

– e-rater uses NLP to identify the features of the faculty-scored essays in its sample collection and store them-with their associated weights-in a database.

– when e-rater evaluates a new essay, it compares its features to those in the database in order to assign a score.

– because e-rater is not doing any actual reading, the validity of its scoring depends on the scoring of the sample essays from which e-rater 's database is created.

Page 20: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

• e-rater FAQs• Q. How often does the computer's score

agree with the score of a faculty reader? – Almost all the time. – ETS researchers found exact agreement, or a

difference of only one point, in as many as 98 percent of the comparisons between the computer's scores and those of a trained essay-reader using the same scoring guides and scoring system.

Page 21: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

• e-rater FAQs• Q. How do students feel about being

scored by a machine? – Most of today's students have had experience with

instant feedback in computer programs and are becoming more comfortable with the idea of computerized scoring.

Page 22: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

• paper– CriterionSM: Online essay evaluation: An application for automated evaluation of

student essays.– authors: Burstein, J., Chodorow, M., & Leacock, C. (2003) – In Proceedings of the Fifteenth Annual Conference on Innovative Applications of

Artificial Intelligence, Acapulco, Mexico. – (This paper received an AAAI Deployed Application Award.) – http://www.ets.org/research/dload/iaai03bursteinj.pdf

• e-rater– trained on 270 essays scored by human readers– evaluates syntactic variety, discourse, topical content, lexical complexity– 50 features

• critique– grammar checker (agreement, verb formation, punctuation, typographical errors)– statistical (bigram) model of English

Page 23: LING 388: Language and Computers Sandiway Fong Lecture 24: 11/17

Interesting things to GoogleQuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

– automated bogus scientific paper generation– http://pdos.csail.mit.edu/scigen/

• SCIgen - An Automatic CS Paper Generator– SCIgen is a program that generates random Computer

Science research papers, including graphs, figures, and citations. It uses a hand-written context-free grammar to form all elements of the papers.

• achievements– one out of two papers got accepted at the World Multiconference on

Systemics, Cybernetics and Informatics (WMSCI)

– Rooter: A Methodology for the Typical Unification of Access Points and Redundancy

• “We implemented our scatter/ gather I/O server in Simula-67, augmented with opportunistically pipelined extensions.”