introducciÓ la major part de les aplicacions interessants de llenguatge requeririen obtenir la...

97
INTRODUCCIÓ • La major part de les aplicacions interessants de llenguatge requeririen obtenir la representació del significat de les oracions.

Upload: brook-henry

Post on 18-Dec-2015

213 views

Category:

Documents


0 download

TRANSCRIPT

INTRODUCCIÓ

• La major part de les aplicacions interessants de llenguatge requeririen obtenir la representació del significat de les oracions.

Estructura predicat argument

La estructura predicat argument descriu les relacions semàntiques que es donen entre les entitats que apareixen en la oració.

-who does what to whom,

-how, where, why?

Estructura predicat argument

I eat sushi

PRED: eat; ARG1: I; ARG2: sushi.

Un exemple més complex.

• En frases complexes tenim més de una proposició.– Mary loves the man who bougth the blue car– P1: Mary loves the man.– P2: The man bought the car.– P3: blue car.

Perquè serveix la estructura sintàgmatica?

• Per obtenir la estructura predicat argument es necessari computar primer la estructura sintagmàtica o al menys una estructura de dependències.

• La estructura sintagmàtica i les corresponents ‘regles’ son les que permeten que el llenguatge sigui composicional.

Però es cert això?

Resum

• Historia (per entendre els objectius de les teories)

• Why phrase structures? (problemes)

• Why dependency grammars? (problemes)

• Why a probabilistic approach?(Brute force vs. theory)

• Estat actual del nostre model i recerca futura.

Una previa:Com millorar els resultats?

• Augmentant el training size?

• Mètodes estadístics més eficients?

• O millorant les teories?

Historia

Grammars as computational theories

Grammars as computational theories

• Cognition is computation.

• A grammar is a form of computation.

Computational theories (Marr 1980)

• What is the goal of the computation?

• Why is it appropriate?

• What is the logic of the strategy by which it can be carried out?

Chomsky’s Goal

A syntactic theory has as a goal to explain the capacity of speakers to judge as acceptable (or ‘generate’) well formed sentences and to rule out ill-formed ones.

Justification

• Syntax is indepedent of semantics.

• Speakers can judge as ill or well-formed new sentences that they have never heard before.

Quin es el origin dels sintagmes?

• No es semàntic. Un NP no es un NP perquè es correspongui amb un argument semàntic.

• Es un NP en base a trets purament sintàctics. Regularitats en la distribució de les paraules en les frases.

• Tests que determinen que es un constituïen (un sintagma) i que no ho és.

Constituency Tests

• “Tests of constituency are basic components of the syntactician’s toolbox. By investigating which strings of words can and cannot be moved, deleted, coordinated or stand in coreference relations, it is possible to draw inferences about the internal structure of sentences.” (Phillips, 1998, p. 1)

• Chomsky assumed that, given the independence of syntax, a theory of syntax can be developed without a semantic theory and ignoring the mapping process, following only the well-formedness goal.

Mapping Goal

A syntactic theory has as a goal to explain the capacity of native speakers to map sentences into the corresponding conceptual representations and vice versa.

 

Mapping Goal

• The mapping goal tries to figure out how linguistic expressions can be mapped in the respective propositional representations in in the most simple and direct way.

Mapping Goal

• (3.a) IBMP gave the company the patent.

• (3.b) PRED: gave; ARG1: IBMP; ARG2: the patent; ARG3: the company.

• (4.a) Low prices.

• (4.b) PRED: low; ARG1: prices.

Well-Formedness Goal

• (3.a) IBMP gave the company the patent.

• (3.b)** IBMP company gave the the patent.

• (4.a) Low prices.

• (4.b)** Prices low

Direct mapping

The carpenter gave the nurse the book.

PRED: gave;

ARG1: the carpenter;

ARG2: the book;

ARG3: the nurse.

El mapping pot ser directe en expresions simples

• Aixo es cert per oracions simples.

• Culicover, Peter W. and Andrzej Nowak. Dynamical Grammar. Volume Two of Foundations of Syntax. Oxford University Press.  2003.

• Roger Schank i col·laboradors en els anys 70.

5941 Mr. NNP - (A0* * * * *5941 Nakamur NNP - *) * * * *5941 cites VBZ cite (V*) * * * *5941 the DT - (A1* * * * *5941 case NN - * * * * *5941 of IN - * * * * *5941 a DT - * (A0* (A0* (A0* (A0* <---------4NLDs5941 custome NN - * *) *) *) *) 5941 who WP - * (R-A0*) (R-A0*) (R-A0*) (R-A0*)5941 wants VBZ want * (V*) * * *5941 to TO - * (A1* * * *5941 build VB build * * (V*) * *5941 a DT - * * (A1* * *5941 giant JJ - * * * * *5941 tourism NN - * * * * *5941 complex NN - * * *) * *5941 in IN - * * (AM-LOC** *5941 Baja NN - * *) *) * *5941 and CC - * * * * *5941 has VBZ - * * * * *5941 been VBN - * * * * *5941 trying VBG try * * * (V*) *5941 for IN - * * * (AM-TMP**5941 eight CD - * * * * *5941 years NNS - * * * *) *5941 to TO - * * * (A1* *5941 get VB get * * * * (V*5941 around IN - * * * * *)5941 Mexican NNP - * * * * (A1*5941 restric NNS - * * * * *5941 on IN - * * * * *5941 foreign JJ - * * * * *5941 ownersh NN - * * * * *5941 of IN - * * * * *5941 beachfr JJ - * * * * *5941 propert NN - *) * * *) *)5941 . . - * * * * *

Direct mapping

• Per Culicover en frases mes complexes no es possible.– Mary loves the man who bougth the blue car– P1: Mary loves the man.– P2: The man bought the car.– P3: blue car.

Direct mapping

• No es possible?– Mary loves the man who bougth the blue car– P1: PRED: loves; ARG1: Mary; ARG2: the man.

– P2: PRED: bought; ARG1:the man; ARG2: the car..

– P3: PRED:blue; ARG1: car.

• Why phrase structures?

• Why dependency grammars?

No son necessaries

• Es pot aconseguir composicionalitat sense computar estructura sintagmàtica

• Es pot fer un mapping directe a la estructura predicat argument sense computar ni una estructura de dependències ni una sintagmàtica.

• Es simplfica considerablement el procés de parsing i el tractament de la ambigüitat.

Temes per poder entrar a fons

• Why phrase structures?

• Why dependency grammars?

• Why a probabilistic approach?

• (al menys la versió “brute-cutre force”)

D-SemMap

V1.0

Vectors and propositions

• A proposition can be represented by a vector of features (Hinton, 1981).

• In order to represent the proposition the vector is divided into “slots”.

• Each element of the proposition is represented in one slot.

Vectors and propositionsVectors and propositionsModule 2Module 2 Semantic classes

Module 1Module 1 POS

SLOT 0 SLOT 1 SLOT 2 SLOT 3 Types & Backs

V MA N DT N

action human artifact entity

SLOT 0 SLOT 1 SLOT 2 SLOT 3 Types & Backs

“Mary drives a bus”

“Mary drives a bus”

Input Word

Slot 0 Slot 1 Slot 2 Slot 3 Type S Back, Test & Subcat.

MODULE 1MODULE 1

Input Layer

Output Layer

Yamada (2003) Nivre (2004)

Magerman (1994)

Ratnaparky (1999)

The carpenter bought a shirt withCredit-card

DT N C

Slot 0 Slot 1 Slot 2 Slot 3

Subcategorization backtracking

V MA PE PA DT N C IIN N C

Output

The

DT

PUT 1

carpenter

N C

PUT 1

bought

V MA PE PA

PUT 0

a

DT

PUT 2PUT 2

shirt

N C

with

IIN

PUT 3

Credit-card

N C

PUT 3

Hidden

Module 2 supervises argument position

P1) PRED: V MA PE PA ( bought)

ARG1: N PR (Mary)

ARG2: DT N C ( a shirt )

ARG3: IIN N C (with pockets)

P1) PRED: get, transfer, give, pay,

ARG1: entity, person

ARG2: entity, object, artifact (shirt)

ARG3: artifact, part-of-dress

MODULE 1MODULE 1 MODULE 2MODULE 2

Subcategorization and Selectional RestrictionsParsing strategy: Attaches first to the current proposition

Binding problem

“Mary bought a shirt with pockets”

P1) PRED: V MA PE PAV MA PE PA ( bought)ARG1: N PRN PR (Mary)ARG2: DT N CDT N C ( a shirt )ARG3: IIN N CIIN N C (with pockets)

P1) PRED: get, transfer, give, pay, get, transfer, give, pay,

ARG1: entity, person entity, person

ARG2: entity, object, artifactentity, object, artifact ARG3: artifactartifact part-of-dresspart-of-dress

MODULE 1MODULE 1 MODULE 2MODULE 2

ARG3: artifactartifact part-of-dresspart-of-dress

BACKCLEARPIZ_IN0

Parsing strategy: Attaches first to the current proposition

“Mary bought a shirt with pockets”

P1:PRED: V MA PE V MA PE ( bought)ARG1: N PRN PR (Mary)ARG2: DT N CDT N C ( a shirt )ARG3:

P2:PRED:ARG1: DTDT N CN C (a shirt)ARG2: IIN N CIIN N C (with pockets)ARG3:

P1:PRED: get, transfer,pay, accept ...

ARG1: entity, person,...

ARG2: entity, object, artifact, shirt ARG3:

P2: PRED: ARG1: entity, object, artifact, shirt ARG2: artifact part-of-dress

ARG3:

MODULE 2MODULE 1

PARSING COMPLEX SENTENCES

Elementary expressions

“a blue shirt”

MODULE 1MODULE 1 MODULE 2MODULE 2

PRED: JJ JJ ( blue )(SLOT 0)

ARG1: DT N CDT N C ( a shirt ) (SLOT 1)

PRED: colour, blue(SLOT 0)

ARG1: entity, object, artifact (SLOT 1)

“the governement`s minister”

ARG1: DT N CDT N C ( the minister ) (SLOT 1)

ARG2: N CN C POSPOS ( goverDTent’s ) (SLOT 2)

TYPE: POSPOS(SLOT type)

ARG1: entity, person...(SLOT 1)

ARG2: entity, person...(SLOT 2)

TYPE: POS(SLOT type)

Complex sentences

• A complex sentence is any sentence that is formed by more than one elementary expression

• A complex sentence requires more than one proposition for its semantic representation

Input Word

Slot 0 Slot 1 Slot 2 Slot 3 Type S Back, Test & Subcat.

MODULE 1MODULE 1

Input Layer

Output Layer

Input Word

COMPLETE SENTENCE STRUCTURE (OPERATIONS WITH VECTORS)

Non Invariant SolutionNon Invariant Solution

Input Layer

Output Layer

Input Word

Slot 0 Slot 1 Slot 2 Slot 3 Type S Back, Test & Subcat.

STACK

Stored Context

Invariant solutionInvariant solution(a kind of shift and reduce parser)(a kind of shift and reduce parser)

Input Layer

Output Layer

Focus of attention (current context)

Input Word

Slot 0 Slot 1 Slot 2 Slot 3 Type S Back, Test & Subcat.

STACK

Stored Context

MODULE 1MODULE 1

Input Layer

Output Layer

Focus of attention (current context)

Modelos concéntricos (Cowan, 1988, 1995, 1999;

Oberauer, 2002)

foco de atención

parte activada de la MLP

memoria alargo plazo(MLP)

A

JA L

retinotopic visual neurons (as found in V1 and V2)

Neurons whose receptive fields are invariant (translation and scale), higher visual areas (inferotemporal cortex)

Covert attention

A

A

JA L

retinotopic visual neurons (as found in V1 and V2)

Neurons whose receptive fields are invariant (translation and scale), higher visual areas (inferotemporal cortex)

Covert attention

A

Attention and invariance

• Shift reduce parser

• Stolke (1990), Sopena (1993), Miikulainen (1996)

Current

Pred:

A1:

A2:

A3:

Flags:

Input Word: The| DTOUTPUT M1: PUT1M2:

The main manager bought some old cars with three wheels.

Generalized Role Labeling using Propositional Representations

Current

Pred:

A1: The

A2:

A3:

Flags: @1

Input Word: The| DTM1: NEXTM2:

The main manager bought some old cars with three wheels.

Generalized Role Labeling using Propositional Representations

Current

Pred:

A1: The

A2:

A3:

Flags: @NEXT @1

Input Word: main | JJ_PRM1: *IZ-INM2:

The main manager bought some old cars with three wheels.

Generalized Role Labeling using Propositional Representations

Current

Pred:

A1:

A2:

A3:

Flags:

Input Word: main| JJ_PRM1: PUT0M2:

The main manager bought some old cars with three wheels.

Top

Pred:

A1: The

A2:

A3:

Flags: @1

Generalized Role Labeling using Propositional Representations

Current

Pred: main

A1:

A2:

A3:

Flags:

Input Word: main| JJ_PRM1: NEXTM2:

The main manager bought some old cars with three wheels.

Top

Pred:

A1: The

A2:

A3:

Flags: @1

Generalized Role Labeling using Propositional Representations

Input Word: manager| DT_NM1: PUT1M2:

The main manager bought some old cars with three wheels.

Current

Pred: main

A1:

A2:

A3:

Flags: @NEXT

Top

Pred:

A1: The

A2:

A3:

Flags: @1

Generalized Role Labeling using Propositional Representations

Input Word: manager| DT_NM1: OZ-OUTM2:

The main manager bought some old cars with three wheels.

Current

Pred: main

A1: manager

A2:

A3:

Flags: @NEXT

Top

Pred:

A1: The

A2:

A3:

Flags: @1

Generalized Role Labeling using Propositional Representations

Input Word: manager| DT_NM1: PUT1M2:

The main manager bought some old cars with three wheels.

Current

Pred:

A1: The

A2:

A3:

Flags: @1 @OZ-OUT

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: manager| DT_NM1: NEXTM2:

The main manager bought some old cars with three wheels.

Current

Pred:

A1: The manager

A2:

A3:

Flags: @1 @OZ-OUT

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: bought| V_MAM1: PUT0M2:

The main manager bought some old cars with three wheels.

Current

Pred:

A1: The manager

A2:

A3:

Flags: @1 @NEXT

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: bought| V_MAM1: NEXTM2:

The main manager bought some old cars with three wheels.

Current

Pred: bought

A1: The manager

A2:

A3:

Flags: @0

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: some| DTM1: PUT2M2:

The main manager bought some old cars with three wheels.

Current

Pred: bought

A1: The manager

A2:

A3:

Flags: @0 @NEXT

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: some| DTM1: NEXTM2:

The main manager bought some old cars with three wheels.

Current

Pred: bought

A1: The manager

A2: some

A3:

Flags: @2

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: old| JJ_PRM1: *IZ-INM2:

The main manager bought some old cars with three wheels.

Current

Pred: bought

A1: The manager

A2: some

A3:

Flags: @2 @NEXT

P:main|A1:manager

Generalized Role Labeling using Propositional Representations

Input Word: old| JJ_PRM1: PUT0M2:

The main manager bought some old cars with three wheels.

Current

Pred:

A1:

A2:

A3:

Flags: @IZ-IN

P:main|A1:manager

Top

Pred: bought

A1: The manager

A2: some

A3:

Flags: @2

Compositionality

• “We now turn to what I think was an important mistake at the core of generative grammar, one that in retrospect lies behind much of the alienation of linguistic theory from the cognitive sciences. Chomsky did demonstrate that language requires a generative system that makes possible an infinite variety of sentences. However, he explicitly assumed, without argument (1965: 16, 17, 75, 198), that generativity is localized in the syntactic component of the grammar” (Jackendoff, 2002)

Compositionality

• The fact that semantics is ‘purely interpretive’ has as a consequence that thought has no ‘independent status’ and it cannot be creative or have an independent capacity of combinatoriality outside of language.

• As Phillips (2004) points out, that thought is purely interpretative and not creative is “a consequence that is likely to be uncomfortable for many, including Jackendoff” (Phillips, 2004 p. 574).

Compositionality

• Semantic (or thought) is “purely interpretative”. Only syntax is creative.

• I think that to believe in ‘absolute free will’ is the main cause of all types of fundamentalism,

Training minimalista

• SS-1-1- (DT The WAIT NEXT)• SS-1-2- (DT_N man PUT1 NEXT)• SS-1-3- (V_MA sold PUT0 NEXT)• SS-1-4- (DT some WAIT NEXT)• SS-1-5- (DT_N offerings PUT2 NEXT)• SS-1-6- (IIN_DT to WAIT NEXT)• SS-1-7- (DT the WAIT NEXT)• SS-1-8- (DT_N president PUT3 NEXT)• SS-1-9- (. . OZ-OUT NEXT)• SS-1-10- (FIN)

Training minimalista(8-10 paraules maxim)

• RL-22-1- (DT a NADA NEXT)• RL-22-2- (DT_N land PUT1 NEXT)• RL-22-3- (CC , PUTtypeCC NEXT)• RL-22-4- (WP2 where TESTARG &NOTEST CLEARmodeCC IZ-IN2

PUTtypeWDT NEXT)• RL-22-5- (DT a NADA NEXT)• RL-22-6- (DT_N saying PUT1 NEXT)• RL-22-7- (V_MA says PUT0 &BACK2 MV23 &BACK_ADJ IZ-INE2

PUTtypeADJ OZ-OUT NEXT)• RL-22-8- (. . OZ-OUT OZ-OUT NEXT)• RL-22-9- (FIN)

Test real, PTBII (55-26)

• s5974: But predictions that central banks of the Group of Seven - G-7 - major industrial nations would continue their massive dollar sales went astray , as the market drove the dollar downward on its own , reacting to Wall Street 's plunge and subsequent price volatility , lower U.S. interest rates and signs of a slowing U.S. economy .

Test

• Ho hem probat amb 254 frases del PTBII

• Els resultats son molt bons. La idea es tenir 0% errors i crec que es pot conseguir.

• Quin es le problema?

NLDs, coordination, comparatives, puntuació

• Dependency grammars and parsers often ignore some classes of dependencies

• Puntuació (guionets, parentesis, comes, dos punts, .....)

NLDs

5941 Mr. NNP - (A0* * * * *5941 Nakamur NNP - *) * * * *5941 cites VBZ cite (V*) * * * *5941 the DT - (A1* * * * *5941 case NN - * * * * *5941 of IN - * * * * *5941 a DT - * (A0* (A0* (A0* (A0* <---------4NLDs5941 custome NN - * *) *) *) *) 5941 who WP - * (R-A0*) (R-A0*) (R-A0*) (R-A0*)5941 wants VBZ want * (V*) * * *5941 to TO - * (A1* * * *5941 build VB build * * (V*) * *5941 a DT - * * (A1* * *5941 giant JJ - * * * * *5941 tourism NN - * * * * *5941 complex NN - * * *) * *5941 in IN - * * (AM-LOC** *5941 Baja NN - * *) *) * *5941 and CC - * * * * *5941 has VBZ - * * * * *5941 been VBN - * * * * *5941 trying VBG try * * * (V*) *5941 for IN - * * * (AM-TMP**5941 eight CD - * * * * *5941 years NNS - * * * *) *5941 to TO - * * * (A1* *5941 get VB get * * * * (V*5941 around IN - * * * * *)5941 Mexican NNP - * * * * (A1*5941 restric NNS - * * * * *5941 on IN - * * * * *5941 foreign JJ - * * * * *5941 ownersh NN - * * * * *5941 of IN - * * * * *5941 beachfr JJ - * * * * *5941 propert NN - *) * * *) *)5941 . . - * * * * *

5961 Fed NNP - * * * *5961 funds NNS - * * * *5961 is VBZ - * * * *5961 the DT - (A1* * * *5961 rate NN - *) * * *5961 banks NNS - (A0*) * * *5961 charge VBP charge (V*) * * *5961 each DT - (A2* * * *5961 other JJ - *) * * *5961 on IN - (A3* * * *5961 overnig JJ - * * * *5961 loans NNS - *) * * *5961 ; : - * * * *5961 the DT - * (A0* (A0* (A0*5961 Fed NNP - * *) *) *)5961 influen VBZ influenc* (V*) * *5961 the DT - * (A1* * *5961 rate NN - * *) * *5961 by IN - * (AM-MNR** *5961 adding VBG add * * (V*) *5961 or CC - * * * *5961 drainin VBG drain * * * (V*)5961 reserve NNS - * * (A1*) (A2*)5961 from IN - * * (AM-MNR*(A1*5961 the DT - * * * *5961 banking NN - * * * *5961 system NN - * *) *) *)5961 . . - * * * *

5920 For IN - * * (AM-PNC * * *5920 the DT - (A0* * * * * *5920 PRI NNP - *) * * * * *5920 to TO - * * * * * *5920 stand VB stand (V*) * * * * *5920 a DT - (A1* * * * * *5920 chance NN - *) * *) * * *5920 , , - * * * * * *5920 Mr. NNP - * * (A0* * * *5920 Salinas NNP - * * *) * * *5920 has VBZ have * (V*) (AM-MOD * * *5920 to TO - * * * * * *5920 press VB press * * (V* * * *5920 on RP - * * *) * * *5920 with IN - * * (A1* * * *5920 an DT - * * * (A0* (A0* (A0*5920 economi JJ - * * * * * *5920 program NN - * * * *) *) *)5920 that WDT - * * * (R-A0*) (R-A0*) (R-A0*) <--------- 3NLDs5920 so RB - * * * (AM-TMP** *5920 far RB - * * * *) * *5920 has VBZ - * * * * * *5920 succeed VBN succeed * * * (V*) * *5920 in IN - * * * * * *5920 lowerin VBG lower * * * * (V*) *5920 inflati NN - * * * * (A1*) *5920 and CC - * * * * * *5920 providi VBG provide * * * * * (V*)5920 moderat JJ - * * * * * (A1*5920 economi JJ - * * * * * *5920 growth NN - * * *) *) * *)5920 . . - * * * * * *

NLDs (Johnson 2002)

• Broad coverage syntactic parsers withgood performance have recently become available (Charniak, Collins), but these typically produce as output a parse tree that only encodes local syntactic information, i.e., a tree that does not include any "empty nodes".

NLDs (Dienes 2003)

• Intuitively, the problem of parsing with NLDs is that the empty elements (EEs) representing these dependencies are not in the input.

• Therefore, the parser has to hypothesise where these EEs might occur – in the worst case, it might end up suggesting exponentially many traces, rendering parsing infeasible (Johnson and Kay1994).

NLDs

• From the point of view of the dependency structure, NLDs are difficult because they violate the assumption that dependency structures are represented as directedtrees. Specifically, NLDs give rise to re-entrancies in the dependency graph, i.e., itis no longer a directed tree but a directed graph, with nodes possibly having multi- ple parents (e.g. apple in Figure 2.2). Now, the parser has to explore a much largersearch space.

NLDs

• Arguably, the search space is much more restricted by an actual grammar that exploits, for instance, the knowledge that buy is a transitive verb and thus requires a direct object.

• Nevertheless, the problem does not disappear. Consider the following example:

 When demand is stronger than suppliers can handle and delivery times lengthen, prices tend to rise. (wsj_0036.mrg)

NLDs

• Non-local dependencies and displacement phenomena have been a central topic of generative linguistics since its inception half a century ago. However …

Many current linguistic theories of non-local dependencies are extremely complex, and would be difficult to apply with the kind of broad coverage described here.

Why a probabilistic approach?

• “Ambiguity and underspecification are ubiquitous in human language utterances, at all levels (lexical, syntactic, semantic, etc.), and how to resolve these ambiguities is a key communicative task for both human and computer natural language understanding” (Manning, 2003).

• At the highest level, the probabilistic approach to natural language understanding is to view the task as trying to learn the probability distribution:

– P(meaning| utterance; context)• A mapping from form to meaning

conditioned by context.

Why a probabilistic approach

quantum mechanics (uncertainty)

Classical Physics (underspecification)

Why a probabilistic approach?

• Collins (1996)– Ambiguity:

• PP-attachment

• Coordination

PP-attachment

• Pierre Vinken, 61 years old, joined the board as a nonexecutive director.

• Pierre Vinken, 61 years old, joined the board as a nonexecutive director.

PP-attachment

• 4-tuple joined board as director

• V = joined, N1 = board, P = as, and N2 = director.

• p (A= l | V=v, N1=n1, P=p, N2=n2)

• p(l | v, n1,p, n2)

Results PP attachment ordered by Accuracy

Average Human Expert (Ratnaparkhi, 1994) 88.20 %

Ratnaparkhi (1994) ID3

77.70 %

Ratnaparkhi (1994) Maximum Entropy Model 81.60 %

Collins and Brooks (1995) Backed-Off Model 84.50 %

Zavrel et al. (1996) Neural Networks

80.00 %

Zavrel et al. (1997) Memory-Based Learning

84.40 %

Yeh and Vilain (1998) Error-driven learning 83.10 %

Committee Machines 1 (Alegre et al, 1999) 86.08 %

Committee Machines 2 (Alegre, 2004) 88.01 %

Method Accuracy

Abney et al. (1999) Boosting

Takahashi (2001) Neural Networks 83.10 %

84.40 %

Krymolowski and Rooth (1998) SNOW 84.80 %

88.01 %

“Mary bought a shirt with pockets”

P1) PRED: V MA PE PAV MA PE PA ( bought)ARG1: N PRN PR (Mary)ARG2: DT N CDT N C ( a shirt )ARG3: IIN N CIIN N C (with pockets)

P1) PRED: get, transfer, give, pay, get, transfer, give, pay,

ARG1: entity, person entity, person

ARG2: entity, object, artifactentity, object, artifact ARG3: artifactartifact part-of-dresspart-of-dress

MODULE 1MODULE 1 MODULE 2MODULE 2

ARG3: artifactartifact part-of-dresspart-of-dress

BACKCLEARPIZ_IN0

Parsing strategy: Attaches first to the current proposition

“Mary bought a shirt with pockets”

P1:PRED: V MA PE V MA PE ( bought)ARG1: N PRN PR (Mary)ARG2: DT N CDT N C ( a shirt )ARG3:

P2:PRED:ARG1: DTDT N CN C (a shirt)ARG2: IIN N CIIN N C (with pockets)ARG3:

P1:PRED: get, transfer,pay, accept ...

ARG1: entity, person,...

ARG2: entity, object, artifact, shirt ARG3:

P2: PRED: ARG1: entity, object, artifact, shirt ARG2: artifact part-of-dress

ARG3:

MODULE 2MODULE 1

Millors resultats

• Alegre (2002)– PTBI 89.8% (PTBII 92.3%)

• Olteanu and Modovan (2005)– PTBII 94%

Parsers de dependencies?

Ambigüitat

• Tota la “artilleria” que cal per resoldre el PP-attachment s’hauria de fer servir per resoldre com col.locar cada paraula de la frase en la estructura?

• En quins fenòmens caldria?

• La resposta es en molt pocs.