mark liberman [email protected] university of pennsylvania · sentiment analysis, machine...

49
What HLT can do for you (and vice versa) Mark Liberman [email protected] University of Pennsylvania

Upload: others

Post on 11-Aug-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

What HLT can do for you (and vice versa)

Mark Liberman

[email protected] University of Pennsylvania

Page 2: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

“Human Language Technology”

HLT: term started in DARPA speech & language programs

= Smart things you can do with text. Also, smart things you can do with speech.

Broader than “Natural Language Processing” (NLP), “Computational Linguistics”, “Speech technology”, . . .

Page 3: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

“Human Language Technology”

Text: Document retrieval, Document classification, Document understanding, Information extraction from text, Summarization, Question answering, Sentiment analysis, Machine translation, . . .

Page 4: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

“Human Language Technology”

Speech: Speech recognition (“speech to text”), Speech synthesis (“text to speech”), Speaker recognition/classification, Diarization (“who spoke when”), Spoken document retrieval, Information extraction from speech, Question answering, Human/computer interaction via speech, Speech-to-speech translation, . . .

Page 5: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

How HLT ought to work (?)

AI Data Human Interaction

Page 6: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

How HLT does work (training phase)

speech,

text,

video,

automatic feature extraction

human annotation: transcription, classification, . . .

statistical model of training data

Page 7: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

How HLT does work (testing phase)

speech,

text,

video,

automatic feature extraction

transcription, classification, . . .

statistical model of training data

Page 8: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

How HLT does work (testing phase)

speech,

text,

video,

independent variables

dependent variables

statistical model of training data

… in some sense, this is just a form of regression …

Page 9: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Decades of Survey/HLT interaction

Ludvic Lebart, Andre Salem, and Lisette Barry, “Exploring Textual Data”, 1998:

… but 15 years later, still not much connection…

Page 10: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Why so little contact?

Maybe HLT didn’t work well enough --

Has there been enough progress to change this?

Maybe the two cultures have been too different?

Has this changed, or will it change?

Are there reasons for increased interaction?

What reasons? What kinds of interaction?

Page 11: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

A key question…

How well does HLT work now? It still depends on: the appropriateness of the statistical model the amount and quality of training data the inherent difficulty of the particular task

So “out of the box” HLT solutions will sometimes work to solve survey-research problems

…but not always, and the learning process may be difficult.

Page 12: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

A side note on human coding

  “Natural annotation” is inconsistent Give annotators a few examples (or a simple definition), turn them loose, and you get:   poor agreement for entities (often F=0.5 or worse)   worse for normalized entities   worse yet for relations among entities

  Why?   Human generalization from examples to principles is variable   Human application of principles is equally variable   NL context raises many hard questions:

… treatment of modifiers, metonymy, hypo- and hypernyms, descriptions, recursion, irrealis contexts, referential vagueness, etc.

  As a result   The “gold standard” is not naturally very golden   The resulting machine learning metrics are noisy

  And F-score of 0.3-0.5 is not an attractive goal!

Page 13: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

The traditional HLT solution

  Iterative refinement of guidelines 1.  Try some annotation 2.  Compare and contrast 3.  Adjudicate and generalize 4.  Go back to 1 and repeat throughout project

(or at least until inter-annotator agreement is adequate)   ~10% (blind) dual annotation throughout task   Convergence is slow (several months per task)   Result is high inter-annotator agreement (>>0.9)   … based on a complex accretion of “common law” (guidelines may be 100s of pages)

  Slow to develop and hard to learn   More consistent than “natural annotation”

Page 14: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

e.g. open-ended response classification / (spoken) document classification

This overlap was featured in that 1998 diagram…

The simplest reason for interaction

Page 15: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

e.g. “sentiment analysis” …

A larger (and better?) set of reasons

Page 16: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

WORD FREQ COEF incredibly 147 1.548 gorgeous 505 1.383 superb 220 1.299 brilliant 328 1.203 beautiful 1374 1.132 wonderful 558 1.081 massive 462 0.9842 wonderfully 456 0.9 beautifully 947 0.8785 opulent 895 0.8475 delicious 3371 0.8332 impressive 1046 0.8008 powerful 1973 0.7809 excellent 1123 0.778 luscious 589 0.7721 huge 706 0.7626 power 1664 0.7222 intense 2324 0.7187 long 5703 0.6714 richness 2346 0.6703

84,158 reviews from Wine Enthusiast magazine 4,153,298 lexical tokens (including punctuation etc.) 29,384 distinct words, of which 5,516 occur 20 or more times 84,158 ratings (numbers between 80 and 100)

Regression of ratings on "bag of words" features:

r=0.864

… i.e. 75% of variance in ratings accounted for

The twenty words with the biggest (significant) coefficients:

“Sentiment analysis”: an easy case

Page 17: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

WORD FREQ COEFF heavy 2406 -0.4882 lean 2120 -0.5332 sour 1614 -0.5527 modest 1164 -0.562 everyday 1074 -0.603 rough 842 -0.6831 rustic 1276 -0.712 simple 3791 -0.7558 short 1273 -0.7701 dimensional 446 -0.8869 thin 1455 -0.933 vegetal 679 -0.9465 watery 398 -0.9947 lack 331 -1.01 lacking 475 -1.069 dull 331 -1.156 lacks 984 -1.234 harsh 449 -1.24 sugary 837 -1.27 acceptable 172 -1.436

84,158 reviews from Wine Enthusiast magazine 4,153,298 lexical tokens (including punctuation etc.) 29,384 distinct words, of which 5,516 occur 20 or more times 84,158 ratings (numbers between 80 and 100)

Regression of ratings on "bag of words" features

r=0.864

.... i.e. 75% of variance in ratings accounted for …

The 20 words with the smallest (significant) coefficients:

Page 18: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions
Page 19: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

An even larger and even better set of reasons

Page 20: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

So what will the future bring?

“Out of the box” HLT solutions will sometimes work to solve survey-research problems

…but not always, and the learning process may be difficult.

The biggest potential benefit is in collaborations where both sides are breaking new ground

How could that work? Thereby hangs a tale . . .

Page 21: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

The story begins in the 1960s…

… with two bad reviews by John Pierce, an executive at Bell Labs who invented the word “transistor” and supervised development of the first communications satellite.

NAS  Responsible  Science  -­‐  7/9/2012   21  

Page 22: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   22  

Page 23: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   23  

In 1966, John Pierce chaired the “Automatic Language Processing Advisory Committee” (ALPAC) which produced a report to the National Academy of Sciences, Language and Machines: Computers in Translation and Linguistics

And in 1969, he wrote a letter to the Journal of the Acoustical Society of America, “Whither Speech Recognition”

Page 24: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

The ALPAC Report

NAS  Responsible  Science  -­‐  7/9/2012   24  

MT in 1966 was not very good, and ALPAC said diplomatically that

“The Committee cannot judge what the total annual expenditure for research and development toward improving translation should be. However, it should be spent hardheadedly toward important, realistic, and relatively short-range goals.”

In fact, U.S. MT funding went essentially to zero for more than 20 years.

The committee felt that science should precede engineering in such cases:

“We see that the computer has opened up to linguists a host of challenges, partial insights, and potentialities. We believe these can be aptly compared with the challenges, problems, and insights of particle physics. Certainly, language is second to no phenomenon in importance. And the tools of computational linguistics are considerably less costly than the multibillion-volt accelerators of particle physics. The new linguistics presents an attractive as well as an extremely important challenge.”  

Page 25: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   25  

John  Pierce’s  views                    about  automaDc  speech  recogniDon                          were  similar  to  his  opinions  about  MT.  

And  his  1969  leKer  to  JASA,                expressing  his  personal  opinion,                      was  much  less  diplomaDc                                than  that  1966  N.A.S.  commiKee  report….  

Page 26: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

“Whither Speech Recognition?”

NAS  Responsible  Science  -­‐  7/9/2012   26  

“… a general phonetic typewriter is simply impossible unless the typewriter has an intelligence and a knowledge of language comparable to those of a native speaker of English.”

“Most recognizers behave, not like scientists, but like mad inventors or untrustworthy engineers. The typical recognizer gets it into his head that he can solve ‘the problem.’ The basis for this is either individual inspiration (the ‘mad inventor’ source of knowledge) or acceptance of untested rules, schemes, or information (the untrustworthy engineer approach). . . .”

“The typical recognizer ... builds or programs an elaborate system that either does very little or flops in an obscure way. A lot of money and time are spent. No simple, clear, sure knowledge is gained. The work has been an experience, not an experiment.”

Page 27: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Tell us what you really think, John

NAS  Responsible  Science  -­‐  7/9/2012   27  

“We are safe in asserting that speech recognition is attractive to money. The attraction is perhaps similar to the attraction of schemes for turning water into gasoline, extracting gold from the sea, curing cancer, or going to the moon. One doesn’t attract thoughtlessly given dollars by means of schemes for cutting the cost of soap by 10%. To sell suckers, one uses deceit and offers glamor.”

“It is clear that glamor and any deceit in the field of speech recognition blind the takers of funds as much as they blind the givers of funds. Thus, we may pity workers whom we cannot respect.”

Page 28: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Fallout from these blasts

NAS  Responsible  Science  -­‐  7/9/2012   28  

The  first  idea:  Try  Ar&ficial  Intelligence  .  .  .  

DARPA  Speech  Understanding  Research  Project  (1972-­‐75)  Used  classical  AI  to  try  to  “understand  what  is  being  said                                                                                                        with  something  of  the  facility  of  a  naDve  speaker”      DARPA  SUR  was  viewed  as  a  failure;  funding  was  cut  off  a_er  three  years  

The  second  idea:  Give  Up.  

                 1975-­‐1986:  No  U.S.  research  funding  for  MT  or  ASR  

Page 29: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   29  

Pierce  was  far  from  the  only  person              with  a  jaundiced  view  of  R&D  investment                    in  the  area  of  human  language  technology.  

By  the  mid  1980s,              many  informed  American  research  managers                        were  equally  skepDcal  about  the  prospects.  

At  the  same  Dme,            many  people  believed  that  HLT  was  needed                        and  in  principle  was  feasible.  

Page 30: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

1985: Should DARPA restart HLT? Charles Wayne -- DARPA program manager – has an idea.

He’ll design a speech recognition research program that –  protects against “glamour and deceit”

•  because there is a well-defined, objective evaluation metric •  applied by a neutral agent (NIST) •  on shared data sets; and

–  and ensures that “simple, clear, sure knowledge is gained” •  because participants must reveal their methods •  to the sponsor and to one another •  at the time that the evaluation results are revealed

In 1986 America,

no other sort of ASR program could have been gotten large-scale government funding.

NAS  Responsible  Science  -­‐  7/9/2012   30  

Page 31: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NIST (Dave Pallett) 1985

NAS  Responsible  Science  7/9/2012   31  

"Performance  Assessment  of  AutomaDc  Speech  Recognizers”,        J.  of  Research  of  the  Na/onal  Bureau  of  Standards:  

DefiniDve  tests  to  fully  characterize  automaDc  speech  recognizer  or  system  performance  cannot  be  specified  at  present.  However,  it  is  possible  to  design  and  conduct  performance  assessment  tests  that  make  use  of  widely  available  speech  data  bases,  use  test  procedures  similar  to  those  used  by  others,  and  that  are  well  documented.  These  tests  provide  valuable  benchmark  data  and  informaDve,  though  limited,  predicDve  power.  By  contrast,  tests  that  make  use  of  speech  data  bases  that  are  not  made  available  to  others  and  for  which  the  test  procedures  and  results  are  poorly  documented  provide  liBle  objec&ve  informa&on  on  system  performance.  

Page 32: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

“Common Task” structure

•  A detailed “evaluation plan” –  developed in consultation with researchers –  and published as the first step in the project.

•  Automatic evaluation software – written and maintained by NIST –  and published at the start of the project.

•  Shared data: – Training and “dev(elopment) test” data

is published at start of project; –  “eval(uation) test” data is withheld

for periodic public evaluations

NAS  Responsible  Science  -­‐  7/9/2012   32  

Page 33: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Not everyone liked it

Many Piercians were skeptical: “You can’t turn water into gasoline,

no matter what you measure.”

Many researchers were disgruntled: “It’s like being in first grade again -- you’re told exactly what to do, and then you’re tested over and over .”

But it worked.

NAS  Responsible  Science  -­‐  7/9/2012   33  

Page 34: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Why did it work? 1.  The obvious: it allowed funding to start (because the project was glamour-and-deceit-proof)

and to continue (because funders could measure progress over time)

2.  Less obvious: it allowed project-internal hill climbing •  because the evaluation metrics were automatic •  and the evaluation code was public This obvious way of working was a new idea to many!

… and researchers who had objected to be tested twice a year began testing themselves every hour…

3.  Even less obvious: it created a culture (because researchers shared methods and results

on shared data with a common metric)

Participation in this culture became so valuable that many research groups joined without funding

NAS  Responsible  Science  -­‐  7/9/2012   34  

Page 35: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

What else it did The common task method created a positive feedback loop.

When everyone’s program has to interpret the same ambiguous evidence, ambiguity resolution becomes a sort of gambling game, which rewards the use of statistical methods.

Given the nature of speech and language, statistical methods need the largest possible training set, which reinforces the value of shared data.

Iterated train-and-test cycles on this gambling game are addictive; they create “simple, clear, sure knowledge”, which motivates participation in the common-task culture.

NAS  Responsible  Science  -­‐  7/9/2012   35  

Page 36: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

The past 25 years Variants of this method

have been applied to many other problems: machine translation, speaker identification, language identification, parsing, sense

disambiguation, information retrieval, information extraction, summarization, question answering, OCR, sentiment analysis, … , etc.

The general experience: 1. Error rates decline by a fixed percentage each year, to an asymptote depending on task and data quality 2. Progress usually comes from many small improvements; a change of 1% can be a reason to break out the champagne. 3. Shared data plays a crucial role – and is re-used in unexpected ways. 4. Glamour and deceit have been avoided.

…and a self-sustaining process was started!

NAS  Responsible  Science  -­‐  7/9/2012   36  

Page 37: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   37  

Page 38: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   38  

Page 39: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   39  

Page 40: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   40  

Page 41: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   41  

Page 42: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   42  

Page 43: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   43  

Page 44: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   44  

Page 45: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

NAS  Responsible  Science  -­‐  7/9/2012   45  

CoNLL  (ACM  SigNLL’s  annual  meeDng):  

Since  1999,  CoNLL  has  included  a  shared  task  in  which  training  and  test  data  is  provided  by  the  organizers  which  allows  parDcipaDng  systems  to  be  evaluated  and  compared  in  a  systemaDc  way.  DescripDons  of  the  parDcipaDng  systems  and  an  evaluaDon  of  their  performances  are  presented  both  at  the  conference  and  in  the  proceedings.  

Page 46: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

The Linguistic Data Consortium

•  Started in 1992 with seed money from DARPA •  LDC has distributed

more than 90,000 copies of more than 1,300 titles to more than 3,200 organizations in more than 70 countries

•  About half of the titles are “common task” datasets –  developed for technology evaluation programs –  released via general catalog after program use

•  ~30 titles added to general catalog per year (current publications queue of about 120 items)

NAS  Responsible  Science  -­‐  7/9/2012   46  

Page 47: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Where we were

NAS  Responsible  Science  -­‐  7/9/2012   47  

ANLP-1983 (First Conference on Applied Natural Language Processing)

34 Presentations:

None use a published data set. None use a formal evaluation metric.

Two examples:

Wendy Lehnert and Steven Shwartz, “EXPLORER: A Natural Language Processing System for Oil Exploration”. Describes problem and system architecture; gives examples of queries and responses. No way to evaluate performance or to compare to other systems/approaches.

Larry Reeker et al., “Specialized Information Extraction: Automatic Chemical Reaction Coding from English Descriptions” Describes problem and system architecture; gives examples of inputs and outputs. No way to evaluate performance or to compare to other systems/approaches.

Page 48: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Where we are

NAS  Responsible  Science  -­‐  7/9/2012   48  

ACL-2010 (48th Annual Meeting of the Association for Computational Linguistics)

274 presentations – Essentially all use published data and published evaluation methods. (A few deal with new data-set creation and/or new evaluation metrics.)

Three examples:

Nils Reiter and Anette Frank, "Identifying Generic Noun Phrases”. Authors are from Heidelberg University; use ACE-2 data.

Shih-Hsiang Lin and Berlin Chen, "A Risk Minimization Framework for Extractive Speech Summarization”. Authors are from National Taiwan University; use Academia Sinica Broadcast News Corpus and the ROUGE metric (developed in DUC summarization track).

Laura Chiticariu et al., ”An Algebraic Approach to Declarative Information Extraction”. Authors are from IBM Research; use ACE NER metric, ACE data, ENRON corpus data.

Page 49: Mark Liberman myl@cis.upenn.edu University of Pennsylvania · Sentiment analysis, Machine translation, . . . “Human Language Technology” ... So “out of the box” HLT solutions

Social science is different… But not that different:

Sharing data and problems –  lowers costs and barriers to entry –  creates intellectual communities –  speeds up replication and extension –  and guards against “glamour and deceit”

(…as well as simple confusion)

NAS  Responsible  Science  -­‐  7/9/2012   49