improving entity linking by modeling latent relations...

12
Improving Entity Linking by Modeling Latent Relations between Mentions Phong Le 1 and Ivan Titov 1,2 1 University of Edinburgh 2 University of Amsterdam 紹介者:駒水 孝裕(名古屋大学) ACL読み会

Upload: others

Post on 06-Oct-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

Improving Entity Linking by Modeling Latent Relations between Mentions

Phong Le1 and Ivan Titov1,21University of Edinburgh 2University of Amsterdam

紹介者:駒水 孝裕(名古屋大学)

ACL読み会

Page 2: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

概要•問題:Entity Linking•従来:ヒューリスティクス,教師あり学習•課題:教師なしでEntity Linkingの実現•手法:

•実験:AIDA-CoNLLデータセットで最新研究と比較

•結果:

1. 関係を隠れ変数として導入2. 複数の関係を考慮することで多様な関係を表現

1. 最新研究よりも精度向上2. 複数の関係の導入が精度に貢献

Page 3: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

Entity Linking• Given: 文書𝐷, 𝐷内の mentions 𝑚# #$%

&

• Find : entities 𝑒# #$%&

1596

World Cup 1966 was held in England …. England won… The final saw England beat West Germany .

coreference beatlocated_in

FIFA_World_CupFIBA_Basketball_

World_Cup...

West_GermanyGermany_national_

football_teamGermany_national_basketball_team

...

EnglandEngland_national_

football_teamEngland_national_basketball_team

...

EnglandEngland_national_football_team

England_national_basketball_team

...

participant_in

Figure 1: Example for NEL, linking each mention to an entity in a KB (e.g. “World Cup” toFIFA WORLD CUP rather than FIBA BASKETBALL WORLD CUP). Note that the first and the sec-ond “England” are in different relations to “World Cup”.

improvement of 0.85% F1 over the best re-ported scores on the standard AIDA-CoNLLdataset (Ganea and Hofmann, 2017). Substan-tial improvements over the relation-agnostic ver-sion show that the induced relations are indeedbeneficial for NEL. Surprisingly its training alsoconverges much faster: training of the full modelrequires ten times shorter wall-clock time thanwhat is needed for estimating the simpler relation-agnostic version. This may suggest that the in-jected structural bias helps to explain regularitiesin the training data, making the optimization taskeasier. We qualitatively examine induced rela-tions. Though we do not observe direct counter-parts of linguistic relations, we, for example, seethat some of the induced relations are closely re-lated to coreference whereas others encode formsof semantic relatedness between the mentions.

2 Background and Related work

2.1 Named entity linking problem

Formally, given a document D containing a list ofmentions m1, ...,mn, an entity linker assigns toeach mi an KB entity ei or predicts that there is nocorresponding entry in the KB (i.e., ei = NILL).

Because a KB can be very large, it is stan-dard to use an heuristic to choose potential can-didates, eliminating options which are highly un-likely. This preprocessing step is called candidateselection. The task of a statistical model is thus re-duced to choosing the best option among a smallerlist of candidates Ci = (ei1, ..., eili). In what fol-lows, we will discuss two classes of approachestackling this problem: local and global modeling.

2.2 Local and global modelsLocal models rely only on local contexts of men-tions and completely ignore interdependencies be-tween the linking decisions in the document (theseinterdependencies are usually referred to as coher-ence). Let ci be a local context of mention mi andΨ(ei, ci) be a local score function. A local modelthen tackles the problem by searching for

e∗i = argmaxei∈Ci

Ψ(ei, ci) (1)

for each i ∈ {1, ..., n} (Bunescu and Pasca, 2006;Lazic et al., 2015; Yamada et al., 2017).

A global model, besides using local contextwithin Ψ(ei, ci), takes into account entity co-herency. It is captured by a coherence score func-tion Φ(E,D):

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci) + Φ(E,D)

where E = (e1, ..., en). The coherence scorefunction, in the simplest form, is a sum overall pairwise scores Φ(ei, ej , D) (Ratinov et al.,2011; Huang et al., 2015; Chisholm and Hachey,2015; Ganea et al., 2016; Guo and Barbosa, 2016;Globerson et al., 2016; Yamada et al., 2016), re-sulting in:

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci)+

!

i =j

Φ(ei, ej , D) (2)

A disadvantage of global models is that exactdecoding (Equation 2) is NP-hard (Wainwrightet al., 2008). Ganea and Hofmann (2017) over-come this using loopy belief propagation (LBP),

mentionentity

1596

World Cup 1966 was held in England …. England won… The final saw England beat West Germany .

coreference beatlocated_in

FIFA_World_CupFIBA_Basketball_

World_Cup...

West_GermanyGermany_national_football_team

Germany_national_basketball_team

...

EnglandEngland_national_football_team

England_national_basketball_team

...

EnglandEngland_national_football_teamEngland_national_basketball_team

...

participant_in

Figure 1: Example for NEL, linking each mention to an entity in a KB (e.g. “World Cup” toFIFA WORLD CUP rather than FIBA BASKETBALL WORLD CUP). Note that the first and the sec-ond “England” are in different relations to “World Cup”.

improvement of 0.85% F1 over the best re-ported scores on the standard AIDA-CoNLLdataset (Ganea and Hofmann, 2017). Substan-tial improvements over the relation-agnostic ver-sion show that the induced relations are indeedbeneficial for NEL. Surprisingly its training alsoconverges much faster: training of the full modelrequires ten times shorter wall-clock time thanwhat is needed for estimating the simpler relation-agnostic version. This may suggest that the in-jected structural bias helps to explain regularitiesin the training data, making the optimization taskeasier. We qualitatively examine induced rela-tions. Though we do not observe direct counter-parts of linguistic relations, we, for example, seethat some of the induced relations are closely re-lated to coreference whereas others encode formsof semantic relatedness between the mentions.

2 Background and Related work

2.1 Named entity linking problem

Formally, given a document D containing a list ofmentions m1, ...,mn, an entity linker assigns toeach mi an KB entity ei or predicts that there is nocorresponding entry in the KB (i.e., ei = NILL).

Because a KB can be very large, it is stan-dard to use an heuristic to choose potential can-didates, eliminating options which are highly un-likely. This preprocessing step is called candidateselection. The task of a statistical model is thus re-duced to choosing the best option among a smallerlist of candidates Ci = (ei1, ..., eili). In what fol-lows, we will discuss two classes of approachestackling this problem: local and global modeling.

2.2 Local and global modelsLocal models rely only on local contexts of men-tions and completely ignore interdependencies be-tween the linking decisions in the document (theseinterdependencies are usually referred to as coher-ence). Let ci be a local context of mention mi andΨ(ei, ci) be a local score function. A local modelthen tackles the problem by searching for

e∗i = argmaxei∈Ci

Ψ(ei, ci) (1)

for each i ∈ {1, ..., n} (Bunescu and Pasca, 2006;Lazic et al., 2015; Yamada et al., 2017).

A global model, besides using local contextwithin Ψ(ei, ci), takes into account entity co-herency. It is captured by a coherence score func-tion Φ(E,D):

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci) + Φ(E,D)

where E = (e1, ..., en). The coherence scorefunction, in the simplest form, is a sum overall pairwise scores Φ(ei, ej , D) (Ratinov et al.,2011; Huang et al., 2015; Chisholm and Hachey,2015; Ganea et al., 2016; Guo and Barbosa, 2016;Globerson et al., 2016; Yamada et al., 2016), re-sulting in:

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci)+

!

i =j

Φ(ei, ej , D) (2)

A disadvantage of global models is that exactdecoding (Equation 2) is NP-hard (Wainwrightet al., 2008). Ganea and Hofmann (2017) over-come this using loopy belief propagation (LBP),

Local model

1596

World Cup 1966 was held in England …. England won… The final saw England beat West Germany .

coreference beatlocated_in

FIFA_World_CupFIBA_Basketball_

World_Cup...

West_GermanyGermany_national_football_team

Germany_national_basketball_team

...

EnglandEngland_national_football_team

England_national_basketball_team

...

EnglandEngland_national_football_team

England_national_basketball_team

...

participant_in

Figure 1: Example for NEL, linking each mention to an entity in a KB (e.g. “World Cup” toFIFA WORLD CUP rather than FIBA BASKETBALL WORLD CUP). Note that the first and the sec-ond “England” are in different relations to “World Cup”.

improvement of 0.85% F1 over the best re-ported scores on the standard AIDA-CoNLLdataset (Ganea and Hofmann, 2017). Substan-tial improvements over the relation-agnostic ver-sion show that the induced relations are indeedbeneficial for NEL. Surprisingly its training alsoconverges much faster: training of the full modelrequires ten times shorter wall-clock time thanwhat is needed for estimating the simpler relation-agnostic version. This may suggest that the in-jected structural bias helps to explain regularitiesin the training data, making the optimization taskeasier. We qualitatively examine induced rela-tions. Though we do not observe direct counter-parts of linguistic relations, we, for example, seethat some of the induced relations are closely re-lated to coreference whereas others encode formsof semantic relatedness between the mentions.

2 Background and Related work

2.1 Named entity linking problem

Formally, given a document D containing a list ofmentions m1, ...,mn, an entity linker assigns toeach mi an KB entity ei or predicts that there is nocorresponding entry in the KB (i.e., ei = NILL).

Because a KB can be very large, it is stan-dard to use an heuristic to choose potential can-didates, eliminating options which are highly un-likely. This preprocessing step is called candidateselection. The task of a statistical model is thus re-duced to choosing the best option among a smallerlist of candidates Ci = (ei1, ..., eili). In what fol-lows, we will discuss two classes of approachestackling this problem: local and global modeling.

2.2 Local and global modelsLocal models rely only on local contexts of men-tions and completely ignore interdependencies be-tween the linking decisions in the document (theseinterdependencies are usually referred to as coher-ence). Let ci be a local context of mention mi andΨ(ei, ci) be a local score function. A local modelthen tackles the problem by searching for

e∗i = argmaxei∈Ci

Ψ(ei, ci) (1)

for each i ∈ {1, ..., n} (Bunescu and Pasca, 2006;Lazic et al., 2015; Yamada et al., 2017).

A global model, besides using local contextwithin Ψ(ei, ci), takes into account entity co-herency. It is captured by a coherence score func-tion Φ(E,D):

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci) + Φ(E,D)

where E = (e1, ..., en). The coherence scorefunction, in the simplest form, is a sum overall pairwise scores Φ(ei, ej , D) (Ratinov et al.,2011; Huang et al., 2015; Chisholm and Hachey,2015; Ganea et al., 2016; Guo and Barbosa, 2016;Globerson et al., 2016; Yamada et al., 2016), re-sulting in:

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci)+

!

i =j

Φ(ei, ej , D) (2)

A disadvantage of global models is that exactdecoding (Equation 2) is NP-hard (Wainwrightet al., 2008). Ganea and Hofmann (2017) over-come this using loopy belief propagation (LBP),

Global model

図は論⽂からの引⽤

Page 4: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

既存手法• mention間の関係• coreference:2つのmentionが同じentityを指示• それ以外の関係

• Cheng and Roth (2013), Ren et al. (2017), など• 人手で作成したルールに基づいてentity間の関係を定義

• 関係抽出自体が難しいタスクなので,ほとんどない

•表現学習• コンテキストの表現とentityの表現の関連性• feature engineeringから発展

• entityに対応するWikipediaのページタイトルの単語ベクトルのコサイン類似度

• 各表現の学習とその間の“単一の”関連性の学習

1596

World Cup 1966 was held in England …. England won… The final saw England beat West Germany .

coreference beatlocated_in

FIFA_World_CupFIBA_Basketball_

World_Cup...

West_GermanyGermany_national_

football_teamGermany_national_basketball_team

...

EnglandEngland_national_

football_teamEngland_national_basketball_team

...

EnglandEngland_national_football_teamEngland_national_basketball_team

...

participant_in

Figure 1: Example for NEL, linking each mention to an entity in a KB (e.g. “World Cup” toFIFA WORLD CUP rather than FIBA BASKETBALL WORLD CUP). Note that the first and the sec-ond “England” are in different relations to “World Cup”.

improvement of 0.85% F1 over the best re-ported scores on the standard AIDA-CoNLLdataset (Ganea and Hofmann, 2017). Substan-tial improvements over the relation-agnostic ver-sion show that the induced relations are indeedbeneficial for NEL. Surprisingly its training alsoconverges much faster: training of the full modelrequires ten times shorter wall-clock time thanwhat is needed for estimating the simpler relation-agnostic version. This may suggest that the in-jected structural bias helps to explain regularitiesin the training data, making the optimization taskeasier. We qualitatively examine induced rela-tions. Though we do not observe direct counter-parts of linguistic relations, we, for example, seethat some of the induced relations are closely re-lated to coreference whereas others encode formsof semantic relatedness between the mentions.

2 Background and Related work

2.1 Named entity linking problem

Formally, given a document D containing a list ofmentions m1, ...,mn, an entity linker assigns toeach mi an KB entity ei or predicts that there is nocorresponding entry in the KB (i.e., ei = NILL).

Because a KB can be very large, it is stan-dard to use an heuristic to choose potential can-didates, eliminating options which are highly un-likely. This preprocessing step is called candidateselection. The task of a statistical model is thus re-duced to choosing the best option among a smallerlist of candidates Ci = (ei1, ..., eili). In what fol-lows, we will discuss two classes of approachestackling this problem: local and global modeling.

2.2 Local and global modelsLocal models rely only on local contexts of men-tions and completely ignore interdependencies be-tween the linking decisions in the document (theseinterdependencies are usually referred to as coher-ence). Let ci be a local context of mention mi andΨ(ei, ci) be a local score function. A local modelthen tackles the problem by searching for

e∗i = argmaxei∈Ci

Ψ(ei, ci) (1)

for each i ∈ {1, ..., n} (Bunescu and Pasca, 2006;Lazic et al., 2015; Yamada et al., 2017).

A global model, besides using local contextwithin Ψ(ei, ci), takes into account entity co-herency. It is captured by a coherence score func-tion Φ(E,D):

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci) + Φ(E,D)

where E = (e1, ..., en). The coherence scorefunction, in the simplest form, is a sum overall pairwise scores Φ(ei, ej , D) (Ratinov et al.,2011; Huang et al., 2015; Chisholm and Hachey,2015; Ganea et al., 2016; Guo and Barbosa, 2016;Globerson et al., 2016; Yamada et al., 2016), re-sulting in:

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci)+

!

i =j

Φ(ei, ej , D) (2)

A disadvantage of global models is that exactdecoding (Equation 2) is NP-hard (Wainwrightet al., 2008). Ganea and Hofmann (2017) over-come this using loopy belief propagation (LBP),

1597

an approximate inference method based on mes-sage passing (Murphy et al., 1999). Globersonet al. (2016) propose a star model which approxi-mates the decoding problem in Equation 2 by ap-proximately decomposing it into n decoding prob-lems, one per each ei.

2.3 Related work

Our work focuses on modeling pairwise scorefunctions Φ and is related to previous approachesin the two following aspects.

Relations between mentionsA relation widely used by NEL systems is corefer-ence: two mentions are coreferent if they refer tothe same entity. Though, as we discussed in Sec-tion 1, other linguistic relations constrain entity as-signments, only a few approaches (e.g., Cheng andRoth (2013); Ren et al. (2017)), exploit any rela-tions other than coreference. We believe that thereason for this is that predicting and selecting rel-evant (often semantic) relations is in itself a chal-lenging problem.

In Cheng and Roth (2013), relations betweenmentions are extracted using a labor-intensive ap-proach, requiring a set of hand-crafted rules and aKB containing relations between entities. This ap-proach is difficult to generalize to languages anddomains which do not have such KBs or the set-tings where no experts are available to design therules. We, in contrast, focus on automating theprocess using representation learning.

Most of these methods relied on relations pre-dicted by external tools, usually a coreference sys-tem. One notable exception is Durrett and Klein(2014): they use a joint model of entity linking andcoreference resolution. Nevertheless their corefer-ence component is still supervised, whereas ourrelations are latent even at training time.

Representation learningHow can we define local score functions Ψ andpairwise score functions Φ? Previous approachesemploy a wide spectrum of techniques.

At one extreme, extensive feature engineeringwas used to define useful features. For example,Ratinov et al. (2011) use cosine similarities be-tween Wikipedia titles and local contexts as a fea-ture when computing the local scores. For pair-wise scores they exploit information about linksbetween Wikipedia pages.

At the other extreme, feature engineering is al-most completely replaced by representation learn-ing. These approaches rely on pretrained embed-dings of words (Mikolov et al., 2013; Penning-ton et al., 2014) and entities (He et al., 2013; Ya-mada et al., 2017; Ganea and Hofmann, 2017) andoften do not use virtually any other hand-craftedfeatures. Ganea and Hofmann (2017) showedthat such an approach can yield SOTA accuracyon a standard benchmark (AIDA-CoNLL dataset).Their local and pairwise score functions are

Ψ(ei, ci) = eTi Bf(ci)

Φ(ei, ej , D) =1

n− 1eTi Rej (3)

where ei, ej ∈ Rd are the embeddings of entityei, ej , B,R ∈ Rd×d are diagonal matrices. Themapping f(ci) applies an attention mechanism tocontext words in ci to obtain a feature representa-tions of context (f(ci) ∈ Rd).

Note that the global component (the pairwisescores) is agnostic to any relations between enti-ties or even to their ordering: it models e1, ..., ensimply as a bag of entities. Our work is in line withGanea and Hofmann (2017) in the sense that fea-ture engineering plays no role in computing localand pair-wise scores. Furthermore, we argue thatpair-wise scores should take into account relationsbetween mentions which are represented by rela-tion embeddings.

3 Multi-relational models

3.1 General form

We assume that there are K latent relations. Eachrelation k is assigned to a mention pair (mi,mj)with a non-negative weight (‘confidence’) αijk.The pairwise score (mi,mj) is computed as aweighted sum of relation-specific pairwise scores(see Figure 2, top):

Φ(ei, ej , D) =K!

k=1

αijkΦk(ei, ej , D)

Φk(ei, ej , D) can be any pairwise score func-tion, but here we adopt the one from Equation 3.Namely, we represent each relation k by a diago-nal matrix Rk ∈ Rd×d, and

Φk(ei, ej , D) = eTi Rkej

Page 5: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

提案手法:multi-relational model•エンティティ間の関連性

• 𝐾個の潜在的な関係の和で表現

1596

World Cup 1966 was held in England …. England won… The final saw England beat West Germany .

coreference beatlocated_in

FIFA_World_CupFIBA_Basketball_

World_Cup...

West_GermanyGermany_national_

football_teamGermany_national_basketball_team

...

EnglandEngland_national_

football_teamEngland_national_basketball_team

...

EnglandEngland_national_football_teamEngland_national_basketball_team

...

participant_in

Figure 1: Example for NEL, linking each mention to an entity in a KB (e.g. “World Cup” toFIFA WORLD CUP rather than FIBA BASKETBALL WORLD CUP). Note that the first and the sec-ond “England” are in different relations to “World Cup”.

improvement of 0.85% F1 over the best re-ported scores on the standard AIDA-CoNLLdataset (Ganea and Hofmann, 2017). Substan-tial improvements over the relation-agnostic ver-sion show that the induced relations are indeedbeneficial for NEL. Surprisingly its training alsoconverges much faster: training of the full modelrequires ten times shorter wall-clock time thanwhat is needed for estimating the simpler relation-agnostic version. This may suggest that the in-jected structural bias helps to explain regularitiesin the training data, making the optimization taskeasier. We qualitatively examine induced rela-tions. Though we do not observe direct counter-parts of linguistic relations, we, for example, seethat some of the induced relations are closely re-lated to coreference whereas others encode formsof semantic relatedness between the mentions.

2 Background and Related work

2.1 Named entity linking problem

Formally, given a document D containing a list ofmentions m1, ...,mn, an entity linker assigns toeach mi an KB entity ei or predicts that there is nocorresponding entry in the KB (i.e., ei = NILL).

Because a KB can be very large, it is stan-dard to use an heuristic to choose potential can-didates, eliminating options which are highly un-likely. This preprocessing step is called candidateselection. The task of a statistical model is thus re-duced to choosing the best option among a smallerlist of candidates Ci = (ei1, ..., eili). In what fol-lows, we will discuss two classes of approachestackling this problem: local and global modeling.

2.2 Local and global modelsLocal models rely only on local contexts of men-tions and completely ignore interdependencies be-tween the linking decisions in the document (theseinterdependencies are usually referred to as coher-ence). Let ci be a local context of mention mi andΨ(ei, ci) be a local score function. A local modelthen tackles the problem by searching for

e∗i = argmaxei∈Ci

Ψ(ei, ci) (1)

for each i ∈ {1, ..., n} (Bunescu and Pasca, 2006;Lazic et al., 2015; Yamada et al., 2017).

A global model, besides using local contextwithin Ψ(ei, ci), takes into account entity co-herency. It is captured by a coherence score func-tion Φ(E,D):

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci) + Φ(E,D)

where E = (e1, ..., en). The coherence scorefunction, in the simplest form, is a sum overall pairwise scores Φ(ei, ej , D) (Ratinov et al.,2011; Huang et al., 2015; Chisholm and Hachey,2015; Ganea et al., 2016; Guo and Barbosa, 2016;Globerson et al., 2016; Yamada et al., 2016), re-sulting in:

E∗ = argmaxE∈C1×...×Cn

n!

i=1

Ψ(ei, ci)+

!

i =j

Φ(ei, ej , D) (2)

A disadvantage of global models is that exactdecoding (Equation 2) is NP-hard (Wainwrightet al., 2008). Ganea and Hofmann (2017) over-come this using loopy belief propagation (LBP),

1597

an approximate inference method based on mes-sage passing (Murphy et al., 1999). Globersonet al. (2016) propose a star model which approxi-mates the decoding problem in Equation 2 by ap-proximately decomposing it into n decoding prob-lems, one per each ei.

2.3 Related work

Our work focuses on modeling pairwise scorefunctions Φ and is related to previous approachesin the two following aspects.

Relations between mentionsA relation widely used by NEL systems is corefer-ence: two mentions are coreferent if they refer tothe same entity. Though, as we discussed in Sec-tion 1, other linguistic relations constrain entity as-signments, only a few approaches (e.g., Cheng andRoth (2013); Ren et al. (2017)), exploit any rela-tions other than coreference. We believe that thereason for this is that predicting and selecting rel-evant (often semantic) relations is in itself a chal-lenging problem.

In Cheng and Roth (2013), relations betweenmentions are extracted using a labor-intensive ap-proach, requiring a set of hand-crafted rules and aKB containing relations between entities. This ap-proach is difficult to generalize to languages anddomains which do not have such KBs or the set-tings where no experts are available to design therules. We, in contrast, focus on automating theprocess using representation learning.

Most of these methods relied on relations pre-dicted by external tools, usually a coreference sys-tem. One notable exception is Durrett and Klein(2014): they use a joint model of entity linking andcoreference resolution. Nevertheless their corefer-ence component is still supervised, whereas ourrelations are latent even at training time.

Representation learningHow can we define local score functions Ψ andpairwise score functions Φ? Previous approachesemploy a wide spectrum of techniques.

At one extreme, extensive feature engineeringwas used to define useful features. For example,Ratinov et al. (2011) use cosine similarities be-tween Wikipedia titles and local contexts as a fea-ture when computing the local scores. For pair-wise scores they exploit information about linksbetween Wikipedia pages.

At the other extreme, feature engineering is al-most completely replaced by representation learn-ing. These approaches rely on pretrained embed-dings of words (Mikolov et al., 2013; Penning-ton et al., 2014) and entities (He et al., 2013; Ya-mada et al., 2017; Ganea and Hofmann, 2017) andoften do not use virtually any other hand-craftedfeatures. Ganea and Hofmann (2017) showedthat such an approach can yield SOTA accuracyon a standard benchmark (AIDA-CoNLL dataset).Their local and pairwise score functions are

Ψ(ei, ci) = eTi Bf(ci)

Φ(ei, ej , D) =1

n− 1eTi Rej (3)

where ei, ej ∈ Rd are the embeddings of entityei, ej , B,R ∈ Rd×d are diagonal matrices. Themapping f(ci) applies an attention mechanism tocontext words in ci to obtain a feature representa-tions of context (f(ci) ∈ Rd).

Note that the global component (the pairwisescores) is agnostic to any relations between enti-ties or even to their ordering: it models e1, ..., ensimply as a bag of entities. Our work is in line withGanea and Hofmann (2017) in the sense that fea-ture engineering plays no role in computing localand pair-wise scores. Furthermore, we argue thatpair-wise scores should take into account relationsbetween mentions which are represented by rela-tion embeddings.

3 Multi-relational models

3.1 General form

We assume that there are K latent relations. Eachrelation k is assigned to a mention pair (mi,mj)with a non-negative weight (‘confidence’) αijk.The pairwise score (mi,mj) is computed as aweighted sum of relation-specific pairwise scores(see Figure 2, top):

Φ(ei, ej , D) =K!

k=1

αijkΦk(ei, ej , D)

Φk(ei, ej , D) can be any pairwise score func-tion, but here we adopt the one from Equation 3.Namely, we represent each relation k by a diago-nal matrix Rk ∈ Rd×d, and

Φk(ei, ej , D) = eTi Rkej

確信度 𝑘番⽬の関連性に基づくスコア

Page 6: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

確信度

• 𝑓(𝑚, 𝑐): mention 𝑚とコンテキスト 𝑐を埋め込む関数

• 𝐃1: ベクトル成分の重みを表現する対角行列• 𝑑: 埋め込むベクトル空間の次元数• 𝑍#41: 正規化項

1598

The weights αijk are normalized scores:

αijk =1

Zijkexp

!fT (mi, ci)Dkf(mj , cj)√

d

"

(4)where Zijk is a normalization factor, f(mi, ci) isa function mapping (mi, ci) onto Rd, and Dk ∈Rd×d is a diagonal matrix.

ei,mi,ci ej,mj,cj

αij1Φ1(ei,ej,D)

αij2Φ2(ei,ej,D)

αij3Φ3(ei,ej,D)

ei,mi,ci ej,mj,cj

(general form)

(rel-norm)

normalize over relations: αij1 + αij2 + αij3 = 1

ei,mi,ci ej,mj,cj

(ment-norm)

normalize over mentions: αi12 + αi22 + … + αij2 + … + αin2 = 1

e1,m1,c1

en,mn,cn

...

...

Figure 2: Multi-relational models: general form(top), rel-norm (middle) and ment-norm (bottom).Each color corresponds to one relation.

In our experiments, we use a single-layer neuralnetwork as f (see Figure 3) where ci is a concate-nation of the average embedding of words in theleft context with the average embedding of wordsin the right context of the mention.1

As αijk is indexed both by mention index jand relation index k, we have two choices forZijk: normalization over relations and normaliza-tion over mentions. We consider both versions ofthe model.

1We also experimented with LSTMs but we could not pre-vent them from severely overfitting, and the results were poor.

3.2 Rel-norm: Relation-wise normalizationFor rel-norm, coefficients αijk are normalizedover relations k, in other words,

Zijk =K#

k′=1

exp

!fT (mi, ci)Dk′f(mj , cj)√

d

"

so that$K

k=1 αijk = 1 (see Figure 2, middle). Wecan also re-write the pairwise scores as

Φ(ei, ej , D) = eTi Rijej (5)

where Rij =$K

k=1 αijkRk.

In foreign policy Bill Clinton ordered U.S. military

tanh, dropout

Figure 3: Function f(mi, ci) is a single-layer neu-ral network, with tanh activation function and alayer of dropout on top.

Intuitively, αijk is the probability of assigning ak-th relation to a mention pair (mi,mj). For ev-ery pair rel-norm uses these probabilities to chooseone relation from the pool and relies on the corre-sponding relation embedding Rk to compute thecompatibility score.

For K = 1 rel-norm reduces (up to a scal-ing factor) to the bag-of-entities model defined inEquation 3.

In principle, instead of relying on the linearcombination of relation embeddings matrices Rk,we could directly predict a context-specific rela-tion embedding Rij = diag{g(mi, ci,mj , cj)}where g is a neural network. However, in prelim-inary experiments we observed that this resultedin overfitting and poor performance. Instead, wechoose to use a small fixed number of relations asa way to constrain the model and improve gener-alization.

3.3 Ment-norm: Mention-wise normalizationWe can also normalize αijk over j:

Zijk =n#

j′=1j′ =i

exp

!fT (mi, ci)Dkf(mj′ , cj′)√

d

"

Page 7: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

埋め込み関数 𝑓(𝑚, 𝑐)

単層ニューラルネットワーク• 活性化関数:tanh• ドロップアウト

1598

The weights αijk are normalized scores:

αijk =1

Zijkexp

!fT (mi, ci)Dkf(mj , cj)√

d

"

(4)where Zijk is a normalization factor, f(mi, ci) isa function mapping (mi, ci) onto Rd, and Dk ∈Rd×d is a diagonal matrix.

ei,mi,ci ej,mj,cj

αij1Φ1(ei,ej,D)

αij2Φ2(ei,ej,D)

αij3Φ3(ei,ej,D)

ei,mi,ci ej,mj,cj

(general form)

(rel-norm)

normalize over relations: αij1 + αij2 + αij3 = 1

ei,mi,ci ej,mj,cj

(ment-norm)

normalize over mentions: αi12 + αi22 + … + αij2 + … + αin2 = 1

e1,m1,c1

en,mn,cn

...

...

Figure 2: Multi-relational models: general form(top), rel-norm (middle) and ment-norm (bottom).Each color corresponds to one relation.

In our experiments, we use a single-layer neuralnetwork as f (see Figure 3) where ci is a concate-nation of the average embedding of words in theleft context with the average embedding of wordsin the right context of the mention.1

As αijk is indexed both by mention index jand relation index k, we have two choices forZijk: normalization over relations and normaliza-tion over mentions. We consider both versions ofthe model.

1We also experimented with LSTMs but we could not pre-vent them from severely overfitting, and the results were poor.

3.2 Rel-norm: Relation-wise normalizationFor rel-norm, coefficients αijk are normalizedover relations k, in other words,

Zijk =K#

k′=1

exp

!fT (mi, ci)Dk′f(mj , cj)√

d

"

so that$K

k=1 αijk = 1 (see Figure 2, middle). Wecan also re-write the pairwise scores as

Φ(ei, ej , D) = eTi Rijej (5)

where Rij =$K

k=1 αijkRk.

In foreign policy Bill Clinton ordered U.S. military

tanh, dropout

Figure 3: Function f(mi, ci) is a single-layer neu-ral network, with tanh activation function and alayer of dropout on top.

Intuitively, αijk is the probability of assigning ak-th relation to a mention pair (mi,mj). For ev-ery pair rel-norm uses these probabilities to chooseone relation from the pool and relies on the corre-sponding relation embedding Rk to compute thecompatibility score.

For K = 1 rel-norm reduces (up to a scal-ing factor) to the bag-of-entities model defined inEquation 3.

In principle, instead of relying on the linearcombination of relation embeddings matrices Rk,we could directly predict a context-specific rela-tion embedding Rij = diag{g(mi, ci,mj , cj)}where g is a neural network. However, in prelim-inary experiments we observed that this resultedin overfitting and poor performance. Instead, wechoose to use a small fixed number of relations asa way to constrain the model and improve gener-alization.

3.3 Ment-norm: Mention-wise normalizationWe can also normalize αijk over j:

Zijk =n#

j′=1j′ =i

exp

!fT (mi, ci)Dkf(mj′ , cj′)√

d

"

図は論⽂からの引⽤

Page 8: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

正規化項

1598

The weights αijk are normalized scores:

αijk =1

Zijkexp

!fT (mi, ci)Dkf(mj , cj)√

d

"

(4)where Zijk is a normalization factor, f(mi, ci) isa function mapping (mi, ci) onto Rd, and Dk ∈Rd×d is a diagonal matrix.

ei,mi,ci ej,mj,cj

αij1Φ1(ei,ej,D)

αij2Φ2(ei,ej,D)

αij3Φ3(ei,ej,D)

ei,mi,ci ej,mj,cj

(general form)

(rel-norm)

normalize over relations: αij1 + αij2 + αij3 = 1

ei,mi,ci ej,mj,cj

(ment-norm)

normalize over mentions: αi12 + αi22 + … + αij2 + … + αin2 = 1

e1,m1,c1

en,mn,cn

...

...

Figure 2: Multi-relational models: general form(top), rel-norm (middle) and ment-norm (bottom).Each color corresponds to one relation.

In our experiments, we use a single-layer neuralnetwork as f (see Figure 3) where ci is a concate-nation of the average embedding of words in theleft context with the average embedding of wordsin the right context of the mention.1

As αijk is indexed both by mention index jand relation index k, we have two choices forZijk: normalization over relations and normaliza-tion over mentions. We consider both versions ofthe model.

1We also experimented with LSTMs but we could not pre-vent them from severely overfitting, and the results were poor.

3.2 Rel-norm: Relation-wise normalizationFor rel-norm, coefficients αijk are normalizedover relations k, in other words,

Zijk =K#

k′=1

exp

!fT (mi, ci)Dk′f(mj , cj)√

d

"

so that$K

k=1 αijk = 1 (see Figure 2, middle). Wecan also re-write the pairwise scores as

Φ(ei, ej , D) = eTi Rijej (5)

where Rij =$K

k=1 αijkRk.

In foreign policy Bill Clinton ordered U.S. military

tanh, dropout

Figure 3: Function f(mi, ci) is a single-layer neu-ral network, with tanh activation function and alayer of dropout on top.

Intuitively, αijk is the probability of assigning ak-th relation to a mention pair (mi,mj). For ev-ery pair rel-norm uses these probabilities to chooseone relation from the pool and relies on the corre-sponding relation embedding Rk to compute thecompatibility score.

For K = 1 rel-norm reduces (up to a scal-ing factor) to the bag-of-entities model defined inEquation 3.

In principle, instead of relying on the linearcombination of relation embeddings matrices Rk,we could directly predict a context-specific rela-tion embedding Rij = diag{g(mi, ci,mj , cj)}where g is a neural network. However, in prelim-inary experiments we observed that this resultedin overfitting and poor performance. Instead, wechoose to use a small fixed number of relations asa way to constrain the model and improve gener-alization.

3.3 Ment-norm: Mention-wise normalizationWe can also normalize αijk over j:

Zijk =n#

j′=1j′ =i

exp

!fT (mi, ci)Dkf(mj′ , cj′)√

d

"

1598

The weights αijk are normalized scores:

αijk =1

Zijkexp

!fT (mi, ci)Dkf(mj , cj)√

d

"

(4)where Zijk is a normalization factor, f(mi, ci) isa function mapping (mi, ci) onto Rd, and Dk ∈Rd×d is a diagonal matrix.

ei,mi,ci ej,mj,cj

αij1Φ1(ei,ej,D)

αij2Φ2(ei,ej,D)

αij3Φ3(ei,ej,D)

ei,mi,ci ej,mj,cj

(general form)

(rel-norm)

normalize over relations: αij1 + αij2 + αij3 = 1

ei,mi,ci ej,mj,cj

(ment-norm)

normalize over mentions: αi12 + αi22 + … + αij2 + … + αin2 = 1

e1,m1,c1

en,mn,cn

...

...

Figure 2: Multi-relational models: general form(top), rel-norm (middle) and ment-norm (bottom).Each color corresponds to one relation.

In our experiments, we use a single-layer neuralnetwork as f (see Figure 3) where ci is a concate-nation of the average embedding of words in theleft context with the average embedding of wordsin the right context of the mention.1

As αijk is indexed both by mention index jand relation index k, we have two choices forZijk: normalization over relations and normaliza-tion over mentions. We consider both versions ofthe model.

1We also experimented with LSTMs but we could not pre-vent them from severely overfitting, and the results were poor.

3.2 Rel-norm: Relation-wise normalizationFor rel-norm, coefficients αijk are normalizedover relations k, in other words,

Zijk =K#

k′=1

exp

!fT (mi, ci)Dk′f(mj , cj)√

d

"

so that$K

k=1 αijk = 1 (see Figure 2, middle). Wecan also re-write the pairwise scores as

Φ(ei, ej , D) = eTi Rijej (5)

where Rij =$K

k=1 αijkRk.

In foreign policy Bill Clinton ordered U.S. military

tanh, dropout

Figure 3: Function f(mi, ci) is a single-layer neu-ral network, with tanh activation function and alayer of dropout on top.

Intuitively, αijk is the probability of assigning ak-th relation to a mention pair (mi,mj). For ev-ery pair rel-norm uses these probabilities to chooseone relation from the pool and relies on the corre-sponding relation embedding Rk to compute thecompatibility score.

For K = 1 rel-norm reduces (up to a scal-ing factor) to the bag-of-entities model defined inEquation 3.

In principle, instead of relying on the linearcombination of relation embeddings matrices Rk,we could directly predict a context-specific rela-tion embedding Rij = diag{g(mi, ci,mj , cj)}where g is a neural network. However, in prelim-inary experiments we observed that this resultedin overfitting and poor performance. Instead, wechoose to use a small fixed number of relations asa way to constrain the model and improve gener-alization.

3.3 Ment-norm: Mention-wise normalizationWe can also normalize αijk over j:

Zijk =n#

j′=1j′ =i

exp

!fT (mi, ci)Dkf(mj′ , cj′)√

d

"

図は論⽂からの引⽤

Page 9: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

正規化方式の選択

1598

The weights αijk are normalized scores:

αijk =1

Zijkexp

!fT (mi, ci)Dkf(mj , cj)√

d

"

(4)where Zijk is a normalization factor, f(mi, ci) isa function mapping (mi, ci) onto Rd, and Dk ∈Rd×d is a diagonal matrix.

ei,mi,ci ej,mj,cj

αij1Φ1(ei,ej,D)

αij2Φ2(ei,ej,D)

αij3Φ3(ei,ej,D)

ei,mi,ci ej,mj,cj

(general form)

(rel-norm)

normalize over relations: αij1 + αij2 + αij3 = 1

ei,mi,ci ej,mj,cj

(ment-norm)

normalize over mentions: αi12 + αi22 + … + αij2 + … + αin2 = 1

e1,m1,c1

en,mn,cn

...

...

Figure 2: Multi-relational models: general form(top), rel-norm (middle) and ment-norm (bottom).Each color corresponds to one relation.

In our experiments, we use a single-layer neuralnetwork as f (see Figure 3) where ci is a concate-nation of the average embedding of words in theleft context with the average embedding of wordsin the right context of the mention.1

As αijk is indexed both by mention index jand relation index k, we have two choices forZijk: normalization over relations and normaliza-tion over mentions. We consider both versions ofthe model.

1We also experimented with LSTMs but we could not pre-vent them from severely overfitting, and the results were poor.

3.2 Rel-norm: Relation-wise normalizationFor rel-norm, coefficients αijk are normalizedover relations k, in other words,

Zijk =K#

k′=1

exp

!fT (mi, ci)Dk′f(mj , cj)√

d

"

so that$K

k=1 αijk = 1 (see Figure 2, middle). Wecan also re-write the pairwise scores as

Φ(ei, ej , D) = eTi Rijej (5)

where Rij =$K

k=1 αijkRk.

In foreign policy Bill Clinton ordered U.S. military

tanh, dropout

Figure 3: Function f(mi, ci) is a single-layer neu-ral network, with tanh activation function and alayer of dropout on top.

Intuitively, αijk is the probability of assigning ak-th relation to a mention pair (mi,mj). For ev-ery pair rel-norm uses these probabilities to chooseone relation from the pool and relies on the corre-sponding relation embedding Rk to compute thecompatibility score.

For K = 1 rel-norm reduces (up to a scal-ing factor) to the bag-of-entities model defined inEquation 3.

In principle, instead of relying on the linearcombination of relation embeddings matrices Rk,we could directly predict a context-specific rela-tion embedding Rij = diag{g(mi, ci,mj , cj)}where g is a neural network. However, in prelim-inary experiments we observed that this resultedin overfitting and poor performance. Instead, wechoose to use a small fixed number of relations asa way to constrain the model and improve gener-alization.

3.3 Ment-norm: Mention-wise normalizationWe can also normalize αijk over j:

Zijk =n#

j′=1j′ =i

exp

!fT (mi, ci)Dkf(mj′ , cj′)√

d

"

関係について正規化

1598

The weights αijk are normalized scores:

αijk =1

Zijkexp

!fT (mi, ci)Dkf(mj , cj)√

d

"

(4)where Zijk is a normalization factor, f(mi, ci) isa function mapping (mi, ci) onto Rd, and Dk ∈Rd×d is a diagonal matrix.

ei,mi,ci ej,mj,cj

αij1Φ1(ei,ej,D)

αij2Φ2(ei,ej,D)

αij3Φ3(ei,ej,D)

ei,mi,ci ej,mj,cj

(general form)

(rel-norm)

normalize over relations: αij1 + αij2 + αij3 = 1

ei,mi,ci ej,mj,cj

(ment-norm)

normalize over mentions: αi12 + αi22 + … + αij2 + … + αin2 = 1

e1,m1,c1

en,mn,cn

...

...

Figure 2: Multi-relational models: general form(top), rel-norm (middle) and ment-norm (bottom).Each color corresponds to one relation.

In our experiments, we use a single-layer neuralnetwork as f (see Figure 3) where ci is a concate-nation of the average embedding of words in theleft context with the average embedding of wordsin the right context of the mention.1

As αijk is indexed both by mention index jand relation index k, we have two choices forZijk: normalization over relations and normaliza-tion over mentions. We consider both versions ofthe model.

1We also experimented with LSTMs but we could not pre-vent them from severely overfitting, and the results were poor.

3.2 Rel-norm: Relation-wise normalizationFor rel-norm, coefficients αijk are normalizedover relations k, in other words,

Zijk =K#

k′=1

exp

!fT (mi, ci)Dk′f(mj , cj)√

d

"

so that$K

k=1 αijk = 1 (see Figure 2, middle). Wecan also re-write the pairwise scores as

Φ(ei, ej , D) = eTi Rijej (5)

where Rij =$K

k=1 αijkRk.

In foreign policy Bill Clinton ordered U.S. military

tanh, dropout

Figure 3: Function f(mi, ci) is a single-layer neu-ral network, with tanh activation function and alayer of dropout on top.

Intuitively, αijk is the probability of assigning ak-th relation to a mention pair (mi,mj). For ev-ery pair rel-norm uses these probabilities to chooseone relation from the pool and relies on the corre-sponding relation embedding Rk to compute thecompatibility score.

For K = 1 rel-norm reduces (up to a scal-ing factor) to the bag-of-entities model defined inEquation 3.

In principle, instead of relying on the linearcombination of relation embeddings matrices Rk,we could directly predict a context-specific rela-tion embedding Rij = diag{g(mi, ci,mj , cj)}where g is a neural network. However, in prelim-inary experiments we observed that this resultedin overfitting and poor performance. Instead, wechoose to use a small fixed number of relations asa way to constrain the model and improve gener-alization.

3.3 Ment-norm: Mention-wise normalizationWe can also normalize αijk over j:

Zijk =n#

j′=1j′ =i

exp

!fT (mi, ci)Dkf(mj′ , cj′)√

d

"

mention について正規化

mention ペアがどの関係に関連するかを表す確率

𝛂 の役割

mention 𝑚# に対して関係𝑘 を通じてmention 𝑚4が関係しているかを表す確率

図は論⽂からの引⽤

𝛂 の役割

Page 10: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

実験• Entity Linkingの精度を評価•比較手法• 提案手法 (rel-norm)• 提案手法 (ment-norm)• 提案手法 (ment-norm (no pad))• 提案手法 (ment-norm (K=1))• 既存手法

•設定• データセット:AIDA-CoNLL• 評価指標:F1のマイクロ平均(5回, 95%信頼区間)• 潜在関連数:6 (rel-norm), 3 (ment-norm)

• 実験的に決定

正規化手法の比較工夫の効果検証複数関連の効果検証

Page 11: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

実験結果

•提案手法 (ment-norm) が最良•計算量的には既存手法より高いが収束が早いので問題にはならない

1601

91% F1 on the dev set, 5 we reduced the learningrate from 10−4 to 10−5. We then stopped the train-ing when F1 was not improved after 20 epochs.We did the same for ment-norm except that thelearning rate was changed at 91.5% F1.

Note that all the hyper-parameters except K andthe turning point for early stopping were set to thevalues used by Ganea and Hofmann (2017). Sys-tematic tuning is expensive though may have fur-ther increased the result of our models.

4.2 Results

Methods Aida-BChisholm and Hachey (2015) 88.7

Guo and Barbosa (2016) 89.0Globerson et al. (2016) 91.0Yamada et al. (2016) 91.5

Ganea and Hofmann (2017) 92.22± 0.14rel-norm 92.41± 0.19

ment-norm 93.07± 0.27ment-norm (K = 1) 92.89± 0.21ment-norm (no pad) 92.37± 0.26

Table 1: F1 scores on AIDA-B (test set).

Table 1 shows micro F1 scores on AIDA-Bof the SOTA methods and ours, which all useWikipedia and YAGO mention-entity index. Toour knowledge, ours are the only (unsupervis-edly) inducing and employing more than one re-lations on this dataset. The others use only onerelation, coreference, which is given by simpleheuristics or supervised third-party resolvers. Allfour our models outperform any previous method,with ment-norm achieving the best results, 0.85%higher than that of Ganea and Hofmann (2017).

Table 2 shows micro F1 scores on 5 out-domaintest sets. Besides ours, only Cheng and Roth(2013) employs several mention relations. Ment-norm achieves the highest F1 scores on MSNBCand ACE2004. On average, ment-norm’s F1 scoreis 0.3% higher than that of Ganea and Hofmann(2017), but 0.2% lower than Guo and Barbosa(2016)’s. It is worth noting that Guo and Barbosa(2016) performs exceptionally well on WIKI, butsubstantially worse than ment-norm on all otherdatasets. Our other three models, however, havelower average F1 scores compared to the best pre-vious model.

The experimental results show that ment-normoutperforms rel-norm, and that mention paddingplays an important role.

5We chose the highest F1 that rel-norm always achievedwithout the learning rate reduction.

4.3 AnalysisMono-relational v.s. multi-relationalFor rel-norm, the mono-relational version (i.e.,Ganea and Hofmann (2017)) is outperformedby the multi-relational one on AIDA-CoNLL,but performs significantly better on all five out-domain datasets. This implies that multi-relationalrel-norm does not generalize well across domains.

For ment-norm, the mono-relational versionperforms worse than the multi-relational one on alltest sets except AQUAINT. We speculate that thisis due to multi-relational ment-norm being lesssensitive to prediction errors. Since it can rely onmultiple factors more easily, a single mistake inassignment is unlikely to have large influence onits predictions.

Oracle

G&H rel-norm ment-norm(K=1)

ment-norm

92

92.5

93

93.5

94

94.5 LBPoracle

Figure 4: F1 on AIDA-B when using LBP and theoracle. G&H is Ganea and Hofmann (2017).

In order to examine learned relations in a moretransparant setting, we consider an idealistic sce-nario where imperfection of LBP, as well as mis-takes in predicting other entities, are taken out ofthe equation using an oracle. This oracle, whenwe make a prediction for mention mi, will tellus the correct entity e∗j for every other mentionsmj , j = i. We also used AIDA-A (developmentset) for selecting the numbers of relations for rel-norm and ment-norm. They are set to 6 and 3,respectively. Figure 4 shows the micro F1 scores.

Surprisingly, the performance of oracle rel-norm is close to that of oracle ment-norm, al-though without using the oracle the differencewas substantial. This suggests that rel-norm ismore sensitive to prediction errors than ment-norm. Ganea and Hofmann (2017), even with thehelp of the oracle, can only perform slightly bet-ter than LBP (i.e. non-oracle) ment-norm. This

ment-norm がrel-norm より良い

複数の潜在関連を取り入れることが良い

図は論⽂からの引⽤

Page 12: Improving Entity Linking by Modeling Latent Relations ...taka-coma.pro/pdfs/ACL2018_reading.pdfACL読み会. 概要 •問題:Entity Linking ... Lazic et al., 2015; Yamada et al.,

まとめ•問題:Entity Linking•従来:ヒューリスティクス,教師あり学習•課題:教師なしでEntity Linkingの実現•手法:

•実験:AIDA-CoNLLデータセットで最新研究と比較

•結果:

1. 関係を隠れ変数として導入2. 複数の関係を考慮することで多様な関係を表現

1. 最新研究よりも精度向上2. 複数の関係の導入が精度に貢献