multi-source uncertain entity resolution: transforming holocaust … · 2016-12-22 · multi-source...

32
Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People Tomer Sagi d , Avigdor Gal e , Omer Barkol d , Ruth Bergman d , Alexander Avram f a Hewlett Packard Labs, Guttwirt Industrial Park, Technion City, Haifa, Israel b Technion - Israel Institute of Technology, Technion City, Haifa, Israel c Yad Vashem, Jerusalem, Israel Preprint submitted to Information Systems December 16, 2016

Upload: others

Post on 06-Jul-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Multi-Source Uncertain Entity Resolution:Transforming Holocaust Victim Reports into People

Tomer Sagid, Avigdor Gale, Omer Barkold, Ruth Bergmand, AlexanderAvramf

aHewlett Packard Labs, Guttwirt Industrial Park, Technion City, Haifa, IsraelbTechnion - Israel Institute of Technology, Technion City, Haifa, Israel

cYad Vashem, Jerusalem, Israel

Preprint submitted to Information Systems December 16, 2016

Page 2: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Multi-Source Uncertain Entity Resolution:Transforming Holocaust Victim Reports into People

Tomer Sagid, Avigdor Gale, Omer Barkold, Ruth Bergmand, AlexanderAvramf

dHewlett Packard Labs, Guttwirt Industrial Park, Technion City, Haifa, IsraeleTechnion - Israel Institute of Technology, Technion City, Haifa, Israel

fYad Vashem, Jerusalem, Israel

Abstract

In this work we present a multi-source uncertain entity resolution model andshow its implementation in a use case of Yad Vashem, the central repository ofHolocaust-era information. The Yad Vashem dataset is unique with respect toclassic entity resolution, by virtue of being both massively multi-source and byrequiring multi-level entity resolution. With today’s abundance of informationsources, this project motivates the use of multi-source resolution on a big-datascale. We instantiate the proposed model using the MFIBlocks entity resolu-tion algorithm and a machine learning approach, based upon decision trees totransform soft clusters into ranked clustering of records, representing possibleentities. An extensive empirical evaluation demonstrates the unique propertiesof this dataset that make it a good candidate for multi-source entity resolution.We conclude with proposing avenues for future research in this realm.

Keywords: Uncertain entity resolution; Blocking; Holocaust

1. Introduction

Cultural heritage institutes are tasked with recording, researching, and pre-serving a culture, often after a severe catastrophe causing a heightened senseof urgency in preserving the records of this culture. These organizations havecollected, over the years, troves of analog artifacts, including films, audio record-5

ings, documents, and pictures. Digitization of these artifacts and extraction ofmetadata and texts, either manually or via OCR techniques, has created adeluge of raw data through which researchers can sift, attempting to create co-herent narratives of a culture now extinct. Recent attempts such as the EHRIproject [5] create infrastructure that enables researchers around the globe to10

access disparate sources of information using a unified interface with underlyingcommon semantics. However, with the growing amounts of data, data inte-gration problems arise. A first step towards recreating the story of a specificperson, community, or place is the unification of all information pertaining to

Preprint submitted to Information Systems December 16, 2016

Page 3: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Figure 1: Guido and Massimo Foa, Cuorgne, 1944

these entities, overcoming different source schemas, languages, political, and15

historical idiosyncrasies.As an example, we bring the story of Guido and Massimo Foa. Figure 1

shows a picture titled “Guido and Massimo Foa, Cuorgne, 1944”. From thepicture we can deduce that there were once a father and son named Guido andMassimo (who is who?) and in 1944 they resided in Cuorgne, Italy. The Yad20

Vashem Names Project,1 has been collecting Holocaust victim reports since1953. Among these are three reports, whose extracted data is presented inTable 1.

BookID First Last Gender DOB Birth Permanent Death Spouse Mother Father

1016196 Guido Foa Male 02/08/36 TorinoItaly

TorinoItaly

Estela Italo

1059654 Guido Foa Male 18/11/20 TorinoItaly

TorinoItaly

Auschwitz Helena Olga Donato

1028769 Guido Foy Male 18/11/20 TurinItaly

CanischioItaly

Olga Donato

Table 1: Three victim reports from the Yad Vashem Names Project DB

Yad Vashem also commemorates non-Jewish individuals who risked theirlives to save Jewish people during the Holocaust. One of those commemorated25

is Clotilde Boggio, who hid a child named Massimo from the Nazis in a villagecalled Cuorgne from 1944 to 1945. Taking into account the information in thesefour sources, a graph such as the one in Figure 2 can be established. During the

1http://www.yadvashem.org/yv/en/remembrance/names/ retrieved June, 2015

3

Page 4: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Figure 2: Knowledge graph of Guido Foa

construction process, data integration decisions need to be taken, e.g., do all ofthe rows in Table 1 refer to the same person? Furthermore, extraction of the30

three records presented in Table 1 is not trivial. A simple query selecting thoserecords whose first name is Guido and last name is Foa would have missed thethird record, which nonetheless contains valuable information.

Weaving information to form narratives, stories told as a sequence of events,has traditionally been a manual process, performed by expert historians. For35

example, Massimo Foa, Guido’s son, grew up to be a historian and wrote a bookdescribing his parents’ story, which enables the validation of the knowledgegraph (Figure 2). The challenge we tackle in this work is to create a robustautomatic procedure to identify and collect all information pertaining to a singleentity from over 500,000 sources in the Yad Vashem database, as a stepping40

stone towards automatically creating narratives for each entity in the database.This data integration challenge can be positioned within the scope of the re-

search area of entity resolution (also known as entity matching, record linkage,and deduplication) [28, 11, 19, 22, 9]. Entity resolution (ER) is at the heartof the data integration problem. The task entails creating a single entity from45

a collection of data records, each revealing some aspects of the entity, with nocommon identifier to rely upon. Examples include identifying accounts belong-ing to the same customer in different operational systems, merging customeraccounts following a bank merger, etc.

Creating narratives from a set of facts poses a new challenge, which is non-50

characteristic of ER applications. Many ER applications require a single crispanswer as the outcome of the process, while here the outcome is a ranked listof possible narratives, which depends on the created ER clusters. Only in rarecases (such as the example given above) one would be lucky enough to find asingle narrative that dominates the others. In most cases, we are faced with55

subjective details of events and based on the context may choose one narrativeover another. To deal with the requirements as were set forward by the Yad

4

Page 5: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Vashem application, we introduce a model for uncertain entity resolution, anER process in which a tuple may be simultaneously associated with multipleentities. With uncertain entity resolution, entities are disambiguated only at60

query time, depending on the query at hand.To support the flexibility uncertain ER requires, we use an entity block-

ing algorithm, MFIBlocks, to create soft blocks (clusters) and apply a machinelearning method using decision trees to transform blocks into ranked associa-tions of records to form entities. We also show how different queries affect the65

ER outcome differently.This paper tells the story of cultural heritage institutions, such as Yad

Vashem, and their effort to collect large amounts of information about the past.However, the multi-source entity resolution task may be relevant to any applica-tion that uses multiple disparate sources of information pertaining to the same70

people and events. While the described project was motivated by a desire topiece together stories from a lost culture, we believe its implications may bene-fit modern applications looking to automatically construct coherent narrativesfrom a multitude of sources.

The work extends the work that was presented at SIGMOD’2016 [? ] in two75

main directions. First, we provide a model for uncertain entity resolution andshow its instantiation using a specific set of algorithms. Second, we significantlyextend our empirical evaluation to show the suitability of the proposed solutionfor the uncertain entity resolution task.

Fix: The rest of the paper is structured as follows. Background on the80

Yad Vashem Names Project is given in Section 2. We then present a model foruncertain entity resolution (Section 3). The algorithmic solution we proposefor uncertain entity resolution is provided in Section 4 followed by details of aconcrete architecture for Yad Vashem in Section 5. We report on an extensiveempirical evaluation of the model (Section 6) and discuss the implications of85

this work and avenues for future research in Section 7.

2. The Yad Vashem Names Project

In 1953 Yad Vashem was established as both a research institute and amemorial. One of its major tasks, beginning in 1954, was the registration ofHolocaust victims’ names on “Pages of Testimony” containing the names and90

biographic details of individual victims.A national campaign of collecting Pages of Testimony in Israel between 1955-

1957 resulted in 800,000 names registered by family members and friends. Col-lection efforts continued: during the 1980’s the average number of incomingPages of Testimony was 14,000-15,000 a year. Following the fall of the Iron95

Curtain and through the 1990’s, the yearly average has risen to 30,000, largelyin Russian. Pages of Testimony are preserved in the Hall of Names.

In September 1991 began the extraction of names and biographic data fromthe Pages of Testimony. This digitization project would later extend to all nameresources in Yad Vashem, including the Archives and the Library. By the end100

5

Page 6: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

of 1998, the Hall of Names digitized 470,000 Pages of Testimony. In addition,500,000 names from major deportation lists were extracted through OCR.

In spring 1999, in agreement with the International Commission of Emi-nent Persons dealing with Swiss dormant bank accounts, Yad Vashem, TadiranSystems Ltd., and Manpower Israel processed the remaining 1.1 million Pages105

of Testimony and scanned the entire collection. In parallel, in April 1999 YadVashem led a renewed campaign of collecting Pages of Testimony resulting in400,000 Pages of Testimony and 50,000 photographs by the end of 2000.

The data was gathered into a database supported by a cataloging and re-trieval system. Efforts have been made to streamline and standardize the data110

and to create advanced retrieval tools. The Central Database of Holocaust Vic-tims’ Names was launched on the Internet in November 2004 and is available inHebrew, English, Russian, Spanish, and German. It registered 14 million visitsin 2014. The database is comprised of name records based on Pages of Testi-mony and archive material such as lists of transports and deportation, inmates115

in camps and ghettos, police registration and property confiscation, as well asmemorial books and commemoration projects. Among them, for example, de-portation lists from Drancy to Auschwitz and Sobibor, card files of inmates inMauthausen, and lists of inmates in ghettos in Transnistria.

The sources are written in Hebrew, Latin, Cyrillic and Greek alphabets, in120

over 30 languages; the greatest part are handwritten, making the process ofdeciphering more complicated. During the registration process, speakers of onelanguage wrote unfamiliar names and places in foreign languages, resulting in avast array of different spellings and semantic variants.

In the extraction process, names and places are registered, barring typing125

mistakes, exactly as they appear in the source to preserve accuracy. Equivalenceclasses of first names, last names and places, as well as professions, personaltitles and family relations, were created to help deal with multiple spellings andvariants. The preprocessing of all misspelling and name synonyms led to a largeyet relatively clean Names project database.130

Figure 3 provides an entity-relationship diagram of the Names Projectdatabase. The central entity is the victim report, assigned a sequential BookIDwhen entered to the database. The database contains 6.5 million victim reportrecords and their auxiliary information such as names and places. Of these, athird was obtained from Pages of Testimony and the others extracted from var-135

ious lists. For testimonies, information regarding the submitter is recorded aswell. However, no unique id exists for a submitter. Thus, the same person mayhave submitted multiple testimonies, about different people (or the same), butthere is no definite method of verifying this automatically without retrieving theoriginal documents and matching handwriting style. Taking a straightforward140

approach and grouping the submitters by first name, last name, and city, re-sults in 514,251 different submitters. Some are obvious duplicates, misspellingsof names and city names, usage of a nickname, or a different transliteration ofthe foreign name, but short of performing entity resolution on the submitterdata, we must remain with this figure. In addition to the testimony submit-145

ters, 16,656 victim lists, gathered from various sources such as transportation

6

Page 7: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Victim ReportBook_ID

Year Birth

SourceID

Description

First Name

Code

Value

Victim's

Preferred

Gender

Last Name

Code

Value

Maiden

Preferred

Place

Code

GPS Coordinates

Preferred

Place Type

Code

Country

CountyRegion

Name

Submitter

ID

First Name

Last NameCountry

City

Address

Father's

Mother's

Mother's

Father's

Marital Status

Code

Description

ProfessionCode

Description

DoD

Relationship to victim

Code

Description

Spouse's

EntityKey Attribute

Description

Entity Other Entity

Entity Other EntityEntity has many other entities

Entity has one other entity

Victim's

Figure 3: Entity Relationship Diagram of the Names Project Database

rid items

1000069 YB 1927,P1 Lubaczow, P2 Lubaczo, P3 Lwow, P4 Poland, F Avraham, L Kesler1000069 P1 Lwow, P2 Lwow, P3 Lwow, P4 Poland, F Avraham, L Apoteker,G 01000069 P1 Antopol, P2 Kobryn, P3 Polesie, P4 Poland, F Yitzhak, F Avram, L Postel, G 01000069 P4 Poland, F Yitzhak, L Postel, G 0

Table 2: Sample records in item-bag form

manifests and concentration camp records, comprise the rest of the sources.

3. Uncertain Entity Resolution: Model

In this section we introduce an uncertain entity resolution model. We startwith an overview of the entity resolution process (Section 3.1), followed by the150

proposed extension of the model to support uncertain entity resolution process(Section 3.2).

3.1. Entity Resolution Overview

Entity resolution is a long standing problem in data integration. Exhaustivealgorithms, comparing all entity pairs in a database, quickly become infeasible155

as the number of pairs grows quadratically with the size of the dataset. There-fore, the prevalent model for entity resolution pipelines (Figure 4) contains ablocking phase in which clusters, groups of records, are created, effectively re-ducing the general Cartesian product of all records to the sum of runs over theCartesian product of records in each group. In practice, blocking techniques160

manage to reduce the number of pair-wise comparisons by 87-97% at the costof a generally acceptable loss in recall. Following the blocking phase, pairwise

7

Page 8: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

comparison is performed using similarity measures (e.g., [12, 10, 29]), machinelearning techniques such as SVM and decision trees (e.g., [26, 27, 4, 8, 25]), orprobabilistic logic.2 These comparisons are then used to classify tuple pairs as165

matched using some predefined threshold. Finally, the pairwise classificationserves as a basis for clustering tuples into entities.

Start Blocking

Pair-wise comparisons

Classification

Raw Record

Set

Blocks of Candidate

pairs

Candidate pairs + Similarity Score

Deterministic DB

Probabilistic DB

Figure 4: Generic ER Process

3.2. Uncertain Entity Resolution Model

The ER process, as described above and as being practically used for manyyears, lacks the ability to accommodate multiple, possible views of a dataset.170

Entity resolution is inherently an uncertain process because the decision to mapa set of tuples to the same entity cannot be made with certainty unless these areidentical in all of their attributes or have a common key. Making deterministicdecisions at various stages of the process may lead to inaccurate results and lossof information [14].175

Several recent works have advocated for the use of probabilistic databasesto represent the multiple views of the outcome of entity resolution (e.g., [2, 3,17]). The essence of this extension to standard ER processing is that pairwisecomparisons can be reasoned about and stored in a probabilistic database, thuseffectively retaining all matching information, and adding a same-as uncertain180

semantic relation between entities. With such models, entities can be resolvedat query time or alternative solutions can be presented, ranked according tosome measure of likelihood.

Uncertain ER is therefore an ER process whose outcome is probabilisticrather than deterministic [14]. The shift towards uncertain ER requires fore-185

2see http://www.umiacs.umd.edu/~getoor/Tutorials/ER_KDD2013.pdf for a tutorial onthese methods.

8

Page 9: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

going crisp clustering and predefined similarity thresholding. We next focus onadapting the blocking and clustering phases to uncertain ER.

The input to the uncertain ER process is a set of tuples T = {t1, · · · , tn}over a schema of k comparable attributes. The output is a set of (possiblyoverlapping) clusters C = {C1, C2, · · · , Cp} (

⋃pi=1 Ci = T ) where each cluster190

represents one entity. In effect, we make the blocking step to be the finalclustering step.

Clusters are created by pairwise tuple comparison using similarity measuressuch as the Jaccard coefficient [22], A tuple pair (ti, tj) is represented as a vectorvi,j = [v1i,j , ..., v

ki,j ]. Each vli,j is a measure of the similarity of the l-th attribute.195

In most cases, the entries of the vector vi,j are in the range [0, 1]. A function cover the values of these entries is used to classify a pair according to a predefinedthreshold.

Beyond tuple pairs, cluster-level constraints are applied, using measures suchas the compact set and the sparse neighborhood [7], to ensure cluster quality.200

The output of the uncertain ER process is a ranked list of results, associat-ing a similarity value for each match, rather than a binary match/ non-matchdecision. We refrain, in this work, from creating a probabilistic distribution overthe participation of tuples in clusters. Rather, we propose a flexible method forranked resolution that enables tuning to the process itself.205

4. Uncertain Entity Resolution: Algorithms

We now introduce the two main algorithmic components of uncertain entityresolution. We use the MFIBlocks algorithm to generate possibly overlapping(soft) record collections (clusters) (Section 4.1). Then, we propose the use ofthe alternating decision tree (ADT) algorithm to transform soft clusters into210

ranked associations that can later serve in performing flexible certainty queryingin Section 4.2.

4.1. Uncertain ER via Soft Clustering

Among the available blocking algorithms that offer soft clusters, clustersthat may share records as an outcome, we have selected MFIBlocks, due to four215

major unique features that make it best suited for uncertain entity resolution.For s detailed literature comparison, we refer the interested reader to [18].

Firstly, MFIBlocks waives the need to manually design a blocking key, thevalue of one or more of a tuple’s attributes. Blocking keys in contemporaryblocking algorithms have to be carefully designed to avoid false negatives by as-220

signing matching tuples to different blocks. Therefore, attributes in the blockingkey should contain few errors and missing values and the design of a blocking keyshould take into account the frequency distribution of values in the attributes ofthe blocking key to balance block sizes. Such careful key design comes with therisk of over fitting the keys to the designer perspective, thus preventing a fair225

evaluation of alternative matching methods. MFIBlocks “lets the data talk” byallowing any combination of attributes to serve as a key, as long as the decisioncan be supported by the data at hand.

9

Page 10: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Second, MFIBlocks localizes the search for similar tuples and is able to un-cover blocks of tuples that are similar in multiple, possibly overlapping sets of230

attributes. MFIBlocks allows a dynamic, automatic, and flexible selection of ablocking key, so that different blocks can be created based on different keys.This approach is in line with the required analysis of uncertain entity resolu-tion, where the current perception of a single-key-fits-all no longer holds. Inparticular, this task should allow multiple levels of granularity, based upon the235

narrative a researcher wishes to follow. For example, in the story of Guido andMassimo Foa, the finest granularity deals with the life of Guido Foa. A differentnarrative involves the whole Foa Family (coarser granularity), while anothermay be interested in all the Jews of Turin.

Blocks created by the algorithm are constrained to satisfy the compact set240

(CS) and sparse neighborhood (SN) [7] properties. As such, local structuralproperties of the dataset are used during the discovery and evaluation of tupleclusters and the number of comparisons for each tuple is kept low, even thoughthe same tuple can appear in several clusters (using multiple keys) simultane-ously. The ability to tune the compact set and sparse neighborhoods properties245

provides flexibility in determining the granularity of the entity resolution pro-cess. In the case of Yad vashem, by allowing a looser compact set setting anddenser neighborhoods, entities can be broadened from a single individual to agranularity of nuclear family and broader social units.

Finally, MFIBlocks is designed to discover entity sets of matching tuples250

with largely varying sizes. MFIBlocks effectively utilizes a-priori knowledge ofthe size of matching entity sets, by discovering clusters of the appropriate sizehaving the largest possible commonality. For the Yad Vashem dataset, archivalexperts estimate that the maximal number of duplicates can be eight recordsor less. This estimate is based upon the source structure of the Names project255

database being mostly first-person testimonials. It was anecdotally supportedby pilot runs with parameter settings that induced much larger blocks but neverproduced valid sets with more than seven records.

For completeness, we next present the MFIBlocks algorithm, starting withMFIs, maximum frequent itemsets. We omit some of the configuration options260

and implementation details for brevity. For a detailed description see [18].

4.1.1. Maximally Frequent Itemsets

Let M = {i1, i2 . . . , ik} be a set of distinct items from which records in adatabase D are built. A record ri ∈ D is composed of a record id (rid) and aset of items Ii ⊆M . Given an item set I ⊆M , the support of I in D is defined265

as the set of records whose item set contains I. I is frequent if the size of itssupport is larger than some threshold minsup. I is maximal (MFI, MaximalFrequent Itemset) if no frequent item-set I ′ exists such that I ⊂ I ′.

Table 2 presents some records from the Yad Vashem dataset after prepro-cessing. Each record is presented as a record id (rid) and a bag of items that270

is created by prefixing a field reference to the value of this field for this record(nulls are omitted). Thus, the item representing the first name Avraham appearsas F Avraham. Given the item set I = {F Yitzhak,L Postel,G 0}, the last two

10

Page 11: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

rows in Table 2 serve as the support of I. minsup = 2 is the highest valuefor which I is a frequent item set. There is no other frequent item set in this275

example that strictly subsumes I, and therefore I is maximally frequent.

4.1.2. The MFIBlocks Algorithm

Let D be a database of n records {ri ∈ D : i ∈ [1, n]}, then the result ofMFIBlocks is a set B of m (possibly overlapping) record blocks {Bj ∈ B : j ∈[1,m]|Bj ⊆ D}.280

Input: Database D; MaxMinSup; pResult: Set of candidate pairs M// Set of Covered Records

1 P = ∅ ;// Set of Candidate Pairs

2 M ← ∅;3 minsup ← MaxMinSup;4 minTh ← 0 ;5 while minsup ≥ 2 ∧P 6= D do

// mine MFI from uncovered records

6 MFIs ← MFI(D \ P ,minsup);7 Blocks ← FindSupport(MFIs);

// Filter blocks

8 Blocks ← Blocks.filter(B → B.size ≤ minsup · p ) ;// Set minTh by NG limitation

9 foreach B ∈ Blocks do10 foreach (ri, rj) ∈ B do11 CandidatePairs ← (ri, rj);12 minTh ← UpdateByNG((ri, rj));

13 end

14 end// filter blocks violating minTh

15 Blocks ← Blocks.filter(B → Score(B) > minTh ) ;16 CandidatePairs ← CandidatePairs.filter(pair → pair.B ∈ Blocks) ;17 P ← P ∪ {ri|∃(ri, rj) ∈ CandidatePairs};18 P ← P ∪ {rj |∃(ri, rj) ∈ CandidatePairs};19 M ←M∪CandidatePairs;20 minsup−− ;

21 endAlgorithm 1: Simplified MFIBlocks Algorithm

Algorithm 1 is supplied with two parameters. MaxMinSup is the maximalvalue of the minsup parameter for record blocks. Thus, if MaxMinSup is setto some arbitrary k, then blocks with a support larger than k are discarded.The second parameter effectively limits the overlap between blocks. The sparseneighborhood condition limits this overlap by capping the number of blocks in285

which a single record may participate.

11

Page 12: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Generating the record blocks is done in two phases. The first mines MFIsfrom the database using, e.g., Borgelt’s FPGrowth algorithm [6] (line 6) andfinds their supporting blocks (line 7). The second phase entails scoring andpruning low-scoring blocks. In line 8, blocks whose size is larger than minsup×p290

are pruned. Lines 9-14 update minTh by finding the minimal block score thatwill prune those blocks violating the sparse-neighborhood condition. Theseblocks are then filtered in line 15 and the pairs they contain are filtered fromthe result set in line 16. This process is done repeatedly, requiring a decreasingminsup for candidate blocks at each iteration and maintaining a set of covered295

records (P) and candidate pairs (M). The algorithm terminates when all recordsare covered or after the minsup parameter is down to one.

4.2. From Soft Clustering to Ranked Resolution

Following the blocking phase, record pairs in the same block are candidatesfor duplication resolution. We argue that applications of uncertain entity resolu-300

tion require flexibility in deciding the extent to which record pairs are resolvedto be duplicates. Such flexibility is accompanied with the need to tune the usageof the entity clusters to the accuracy required for different applications. For ex-ample, in the Yad Vashem use-case, a user app relaying historical information,including the number of people perished in the Holocaust in various parts of305

Europe, requires a single deterministic answer. On the other hand, a personsearching for perished relatives can control the size of the response by tuninga certainty parameter in a Web-query interface. Therefore, the output of theentity resolution is a ranked list of results, associating a similarity value for eachmatch, rather than a binary match/ non-match decision.310

To create a ranked resolution we use a classifier that learns the appropriatesimilarity value. When dealing with multi-source entity resolution the data maybe sparse , which causes concern when choosing a classifier. For example, onerecord in the Yad Vashem repository may contain only a first name, a last nameand a birth year while another record may contain first, last and father’s names315

without a birth year. Thus, we are required to use techniques that are robustto disparity between record attributes.

We chose the alternating decision trees (ADTrees) classifier [13], which sup-plies a single decision tree, easily interpretable, yet robust to variability in bothtraining and testing data. Furthermore, the technique provides each match with320

a prediction score, allowing ranking of the results and tuning accuracy versusnumber of returned results.

For exposition completeness we briefly describe ADTrees. Consider the ex-amples in figures 5 and 6. Figure 5(a) contains a standard decision tree. Thetree includes two decision nodes and three prediction leaves. The tree performs325

a binary classification of (a, b) ∈ R2 into {−1,+1}. The example in Figure5(b) implements the same binary classification using an alternating decisiontree (ADTree). It consists of two types of nodes: Splitter nodes, marked withrectangles, which act as decision nodes, and prediction nodes (in elipses), eachof which holds a real valued number that can be semantically understood as the330

confidence gain that a state adds to the classification. The value in a leaf node

12

Page 13: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Figure 5: Decision tree standard representation (a) and as an ADTree (b) [13]

is computed as the sign of the sum of all prediction nodes on the path from theroot to the leaf. This way, in the example, for (a, b) = (3.9, 0.9) the value wouldbe sign(+0.5− 0.7− 0.2) = −1.

Figure 6: A general alternating tree [13]

General alternating decision trees allow each splitter node to split more than335

once. Semantically, this allows to create a Union phrase for different decisioncriteria. The resulting score would be the sum of the prediction nodes in thetraversed sub-tree that is spanned from the root to all the accepted leaves.In Figure 6, an additional splitter child node was added to the two internalprediction nodes. In this case, the input (a, b) = (3.9, 0.9) would result in three340

13

Page 14: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

leafs, and its value would be sign(+0.5 + 0.3− 0.7− 0.1− 0.2) = −1.The constructed ADTree is much denser comparing to standard decision

trees, which contributes immensely to the ease with which it can be understood.An additional feature, which we exploit, is the ability to disregard the sign

operation and use the resulting score as a confidence score with respect to the345

classification at hand. This score serves as the basis of a ranked decision insteadof a deterministic classification. Finally, this method allows graceful handlingof missing values, as in such cases the computation considers only reachabledecision nodes [13]. In many cases, accuracy will not be degraded significantly,which is a crucial property when regarding our multi-source schema-diverse350

dataset.

5. Multi-source Entity Resolution in the Names Project

We now present the details of the instantiation of uncertain entity resolutionin the Names Project, allowing for multiple levels of resolution and incorporatinginformation about multiple sources.355

5.1. Data Preparation

From the data described above, two subsets were extracted, of which onewas tagged.

ItalySet A homogeneous dataset comprised of all records having Italy as thevictim’s place of residence. This Dataset contains 9,499 tagged records360

and is available for public use.3

RandomSet A stratified random sample of the full dataset. Six geographicalregions were selected from the dataset, each representing a different pre-Holocaust Jewish community. Differences were either cultural-linguisticor in the progression of persecution during WWII itself. This dataset365

contains 100,000 records.

Each field in the original data was given a unique prefix, which was addedto the items. Thus, when the literal ‘Moshe’ appears in the first name field of arecord, it is entered in that record’s item bag as ‘FN Moshe’. A person may havemultiple occurrences in some attributes, such as first name, and war-time place.370

This is supported by the bag-of-items model used by the MFIBlocks algorithmand the items are added independently.

To obtain expert tags, MFIBlocks was run several times and with several con-figurations on the Italy set. The candidate pairs from this process were bundledinto a tagging application (see screen-shot from user manual in Figure 7). The375

application presents users with the candidate matches sorted by similarity whilehighlighting the differences between the records. Yad Vashem archival expertstagged each pair with one of the following {Yes, Probably Yes, Maybe, Probably

3https://github.com/tomersagi/yv-er

14

Page 15: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Write your comments here

Yellow: only some of the names match

Decide on the match here

Use the navigation buttons to scroll between records

Results are ordered by descending similarity score

Press here to jump to this record

Figure 7: Screen-shot from Tagging Application User Guide

No, No}. A Maybe tag indicates that the information contained in the pair isinsufficient to decide whether they constitute a matched pair. Figure 8 presents380

an analysis of the expert tags assigned to candidate pairs. We examined theproportion of pairs assigned to each tag for similarity bins from 0.1 to 1.0. Theanalysis focused on aberrations such as low scoring pairs assigned a Yes tag orhigh scoring pairs assigned a No tag. These were used to validate tagging andtweak our similarity function.385

Following the qualitative analysis, the tagged set was simplified by combiningYes with Probably Yes and No with Probably No. The revised tagged set wasused to both train the classifier and experiment with different configurations ofthe system (See Section 6). Since the exhaustive set of pairs (45 × 106 for theItalian set) is too large for humans to review, we concede that there may well be390

additional matched pairs not found by any configuration and therefore remainuntagged (false negatives). In addition, we invite future researchers to carefullyexamine any false positives reported by the algorithms, as these may containpairs that were untagged by omission. We welcome submissions of such suspectpairs to the correspondent author from Yad Vashem for validation.395

Finally, to facilitate the ADT algorithm, 48 similarity features over matchedrecord pairs were defined. We constructed every conceivable similarity featuregiven the record attributes, assuming these will be pruned by the ADT al-gorithm. The algorithm constructs the most efficient decision tree given thetraining data, and omits any similarity features that do not contribute to ac-400

15

Page 16: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

0%

10%

20%

30%

40%

50%

60%

70%

80%

90%

100%

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Tag

pro

po

rtio

n

Similarity

No

Probably-No

Maybe

Probably Yes

Yes

Is Match?

Figure 8: Tag-Similarity Comparison

curacy. For each feature, one of three types of comparisons was applied. First,categorical attributes, such as name, gender and profession, are compared usinga binary similarity feature. Second, the trinary option was applied to attributevalues with a small number of words, those for which less than 5% of the recordshave more than two words. Finally, for the non-discrete attributes, a normalized405

distance was used, using Yad Vashem expert opinion to tune the normalizationfactor. The final list of features is as follows. Due to the unique properties ofADT, if a record does not have a value for some attribute, then that feature isnot used for the matched pair.sameXName For each of the seven attributes (First, Last, Spouse, Father,410

Mother, Mother’s Maiden, Maiden) the corresponding feature is valued yeswhen all of the matched pairs’ names of this type were the same, partialwhen only some were the same and no if none matched. For example,comparing a record with first names {John, Harris} with another recordwhose first name is John would result in partial.415

XnameDist For each of the seven names, a corresponding feature calculates theJaccard similarity (a value of 1.0 is perfectly similar) between the cor-responding names and taking the maximum over multiple names. Thisfeature is included since we encountered some cases of clerical errors(Bella→Della) even though extensive preprocessing by Yad Vashem ex-420

perts should preclude most typos, spelling mistakes, and multiple spellingand pronunciation induced versions of a name.

BXDist For each of the birth date components (day, month, and year) this fea-

16

Page 17: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

PreprocessingRecord Item Sets

ItemsMFI Blocks

Matched PairCandidates

ADT model

Frequent Item Mining using FP-Growth

Record Block Construction and Pruning using Spark

Figure 9: System Architecture

ture measures the distance between the values, normalized by a maximaldistance (31 for days, 12 for months, 100 for years).425

samePlaceXPartY For each of the four place types (Birth, Permanent, War-time,Death), and for each of the four place components (City, County, Region,Country) a feature is valued yes if the place part is the same.

PlaceXGeoDistance For each of the four place types, this feature represents thedistance in KM between the same place type in the two compared records.430

For example, for two records with birth places of Turin and Moncalieri, thevalue would be 9 (KM).

sameSource This feature is valued true if the records are from the same list orare pages of testimony submitted by the same person.

sameGender This features is valued true if the records are of the same gender.435

sameProfession This features is valued true if the records have the same pro-fession code.

Schema reconciliation of the various sources was performed continuously overthe 60 years in which the data was collected. We have reasonable confidence inthe semantics of the different place attributes (Birth, Death, Permanent, War-440

time). Therefore, we do not compare place values across attributes (e.g., Birthto Permanent). The high accuracy result (∼ 95%, see Section 6.4) obtained bythe ADT algorithm, choosing only 8-10 features of the 48 defined, allowed us toconclude the feature evaluation process at a relatively early stage of the project.

5.2. System Architecture445

The system architecture is presented in Figure 9. A preprocessing stepconverts the records to a collection of pairs (record id, item-set) and creates anindex that maps each item to the list of records in which it appears. The datasetis then fed to the MFIBlocks algorithm using Borgelt’s FP Growth [6] algorithmto mine frequent items and an Apache Spark-based distributed algorithm to450

process the MFIs and create blocks. Finally, a trained ADT model is used forclassification and similarity calculation.

17

Page 18: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

(Prior) -0.29

sameFatherName

mfNameDist<0.73

no:+

1.53

yes:-0.72

no:-1.3

ffNameDist<0.47no

:-0.2

5

yes:-0.86

no:

-1.3

yes:+0.54

Figure 10: Fragment of Final ADT tree

The MFIBlocks algorithm was modified for the purpose of this project asfollows. The original MFI algorithm assumed a collection of q-grams in therecords’ item-sets. However, in this dataset, dates and geo locations supplant455

textual data such as names and places. Therefore, we experimented with replac-ing the union of Jaccard Similarity between item q-grams used to score blocks inthe original work with a custom item similarity score defined as follows (Eq. 1).

fsim(i1, i2) =

0 type(i1) 6= type(i2)

jw(i1, i2) type(i1) = Name

1− |i1−i2|50

type(i1) = Y ear

1− monthDiff(i1,i2)12

type(i1) = Month

1− dayDiff(i1,i2)31

type(i1) = Day

Max(0, 1− geoDist(i1,i2)100

) type(i1) = Geo

(1)

Using the similarity features described above, the ADT tree was trainedand accessed at run-time using Weka 3.6 [16]. A fragment of the final model460

is described in Figure 10. For each candidate record pair, all branches whoseconditions match the features of this pair are traversed and their value accu-mulated to the overall similarity score. For example, a pair that has differentfather names, whose distance between father names is 0.2, and no mother firstname in one of the records would score −1.3 + −0.25 = −1.55. To calculate465

accuracy, we use the default classification of values lower than or equal to zeroto non-match and above zero to match. However, as previously discussed, thereis merit in retaining the similarity score as well. The full tree under differentexperimental conditions is described in Section 6.5, and presented in tables 7and 8.470

18

Page 19: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

6. Empirical Evaluation

In this section we present evaluations performed on the Yad Vashem datasetand its subsets described in Section 2. We begin with the experiment setup (Sec-tion 6.1), an evaluation of the data pattern variance in the dataset (Section 6.2),and a performance oriented evaluation of the system (Section 6.3). Evaluation475

of the ADT classifier using different feature sets is given in Section 6.4. Weevaluate the full system on the Italian subset in Section 6.5, and compare toalternative state-of-the-art algorithms in Section 6.6.

6.1. Set-up

Classifier training and evaluation was done on an HP Ultrabook laptop with480

4 cores and 8GB RAM. MFIBlocks was run on an HP DL380 server with 24 coresand 64GB RAM of which 30GB were allocated to the JVM heap and 30GB toan Apache Spark 1.2.1 pseudo cluster. The single-threaded implementationof FPGrowth [6] was used by MFIBlocks to mine frequent item sets and wasthe most efficient known sequential version of this algorithm at the time of485

publication. However, as the following evaluation shows, the sequential natureof the algorithm caused it to be the bottleneck of the system as it could onlyutilize one of the 24 cores available on the server and accounted for 90% of theruntime.

6.2. Data Patterns490

18,567

5,318

1,897

490 960

2,000

4,000

6,000

8,000

10,000

12,000

14,000

16,000

18,000

20,000

0

0.5

1

1.5

2

2.5

3

3.5

4

4.5

10 100 1000 10000 more

Co

un

t #

pat

tern

s

Sum

of

#rec

ord

sM

illio

ns

# Records with this pattern

count Sum

Figure 11: Data Pattern Counts

An important ramification of the multi-source nature of this dataset is theexpected schema variability between sources. In order to examine this variabil-ity, we performed a pattern count over the entire dataset, where a pattern is a

19

Page 20: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

set of item types and records share a type if they have values assigned for thesame item types. Figure 11 provides an anlysis of the results. The x-axis groups495

records based on the number of records that share a pattern. The leftmost valueof 10 represents patterns that are shared by 10 records or less. The next valueof 100 represents patterns that are share by more than 10 records and up to 100records, etc. The bar graph shows how many such patterns exist in the datasetwhile the line graph shows how many records participate in such patterns.500

According to Figure 11, there are 96 patterns that each is shared by morethat 10,000 records with a total of over four million such records. In this group,the most prevalent pattern is shared by half a million records and contains onlyFirstName, LastName, Gender, and Permanent Place. 18,567 different patternshave less than ten records each. The full information pattern containing all505

possible item types is shared by only 40,191 records.

Item TypeFull Set 10K Italy Set 100K Set

Records % Records % Records %

Last Name 6,478,181 98% 9,402 99% 96,656 97%First Name 6,352,481 97% 9,402 99% 97,527 98%Gender 5,778,172 88% 9,223 97% 86,474 86%DOB 4,185,298 64% 6,345 67% 65,005 65%Father’s Name 3,452,056 52% 7,392 78% 49,019 49%Mother’s Name 2,639,497 40% 5,583 59% 39,816 40%Spouse Name 1,759,845 27% 2,022 21% 25,964 26%Maiden Name 819,618 12% 1,201 13% 12,399 12%Mother’s Maiden 758,303 12% 1,954 13% 11,950 12%Permanent Place 4,634,209 70% 8,348 88% 83,308 83%Wartime Place 3,841,196 58% 6,806 72% 70,009 70%Birth Place 2,344,618 36% 8,563 90% 58,399 58%Death Place 2,253,248 34% 5,675 60% 42,541 43%Profession 2,332,427 35% 2,559 27% 27,209 27%

Table 3: Item Type Prevalence

Table 3 report on the prevalence of item types in the full dataset and inthe public subsets provided. As expected, prevalence of the stratified randomsample of 100,000 records is very similar to that of the full dataset. Notabledifferences relate to higher prevalence of Place information since it was specifi-510

cally used to create the stratified sample. The homogeneity of the Italy sampleis a probable antecedent to its different data pattern. Notably, a Person’s fathername was a major part of their identity in this community and, thus, appearsin most records.

In addition to prevalence, we examined the cardinality of each item type.515

Low cardinality items have the potential to be good markers for matched recordpairs and be used as the basis for entity clusters, while high cardinality itemsare often discarded by the blocking algorithm, although may still be used bythe classifier. For example, using Gender as the basis for a block will contribute

20

Page 21: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Item Type10K Italy Set 100K Set

Items Records/Item Items Records/ItemLast Name 1,495 6 20,541 5First Name 775 12 1,544 63Gender 2 4,612 2 43,237Maiden Name 505 2 5,019 2Mother’s Maiden Name 496 4 4,731 3Mother’s First Name 382 15 825 48Profession 355 7 2,115 13Spouse Name 430 5 955 27Father’s Name 372 20 777 63Birth Day 31 127 31 738Birth Month 12 332 12 2,068Birth Year 98 64 119 541Birth City 670 12 9,519 6Birth County 278 19 1,112 44Birth Region 91 86 151 223Birth Country 33 259 52 1,116War City 290 23 6,976 8War County 114 30 1,715 25War Region 54 116 195 218War Country 20 338 45 1,556Perm. City 226 36 8,328 10Perm. County 73 70 1,051 73Perm. Region 19 425 81 800Perm. Country 1 8,348 23 3,622Death City 209 26 3,802 11Death County 87 7 924 13Death Region 53 19 238 80Death Country 18 310 45 945

Table 4: Item Type Cardinality

little to the block’s quality. By using the MFIBlocks algorithm we ensure high520

cardinality items are not used as block keys. However, these items are availablefor the classifier. In this case, Gender may disqualify candidate pairs of differ-ent genders. Table 4 presents the cardinality and average number of recordsper item for the dataset items on the two sample subsets. One should notethat even if a field has high cardinality, the extremely diverse data patterns in525

this dataset significantly lower the probability of records sharing many high-cardinality items.

6.3. Performance

In our second evaluation we examine how the system’s runtime scales withdataset size and the minsup parameter. We employ the method reported in [18]530

21

Page 22: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

to prune the .03% most frequent items and compare the runtime with and with-out pruning. Figure 12 shows the runtime in seconds, plotted at log scale againstthe minsup parameter value for four test conditions. Series 6.5M indicates thefull dataset of 6.5 million records, 600K is a random sample of 600,000 records.

1

10

100

1000

10000

100000

2 2.5 3 3.5 4 4.5 5

log(

Ru

nti

me

(sec

))

Min-sup

FPGrowth Run-Time

6.5M

6.5M,Prune

600K

600K,Prune

Figure 12: Run-time comparison

Results indicate that runtime increases exponentially as minsup decreases,535

and linearly with dataset size. Detailed analysis of runtime revealed that theFPGrowth implementation, being sequential, was the performance bottleneckof this set-up as it used only a single core and could not be trivially parallelizedusing the Spark cluster.

6.4. ADT Classifier540

Next, we examine the accuracy of the ADT classifier over the candidatepairs returned by the MFIBlocks algorithm under various conditions. Recallthat the golden standard is tagged by three tag types: Yes, No, and Maybe.

Condition N AccuracyMaybe:=No 10,016 94.2%Maybe values omitted 9,406 96.4%Identify Maybe values 10,016 95.1%

Table 5: Classifier Quality - Maybe values

The semantics of the Maybetagged pairs is that they do545

not carry enough informationto be correctly classified. Whentraining the classifier, shouldwe keep such pairs as a distinctgroup to be identified at run-550

time as well? Do we classifythese as non-matches? Or do we omit these examples? It should be notedthat of the 10,017 tagged pairs, 611 were tagged as Maybe limiting the effect onaccuracy to a 6% difference. As demonstrated in Table 5, accuracy levels are

22

Page 23: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

stable around 95% in all configurations, with a slight advantage to the model555

trained on a set without Maybe values.We attempt to avoid over-fitting of the ADT model by removing a unique

source to the Italian dataset. This person, which we will refer to by his initialsMV, supplied 1,400 of the 9,499 records in the Italy subset, which is unusual byany standard, and uncommon in the general Yad Vashem dataset, where most560

submitters submit 1-5 testimony pages. Furthermore, MV submitted a uniquedata pattern for all his submissions, namely {FirstName, LastName, FatherName,BirthPlace, Death Place}. While this submitter provides valuable information,including these reports in the classification model runs the risk of over fitting themodel to the Italian subset, since the phenomenon of a single person submitting565

over a 1,000 pages of testimony is extremely rare and occurs only twice morein 6.5 million records. Of the 10,016 tagged pairs, 3,183 are pairs involving anMV record and so the reduced dataset is comprised of 6,833 tagged pairs.

Condition N AccuracyWith MV 9,406 96.5%Without MV 6,833 94.2%

Table 6: Classifier Quality - MV source

As Table 6 shows, model accuracydrops by 2.3% when MV reports are570

removed. However, comparing thetwo models (tables 7 and 8), we ob-serve that the MV-less model puts lessemphasis on father name (FFN) andhigher emphasis on same first name,575

which may better represent the behavior of a general subset of this dataset thanthe MV records.

: -0.289— (1)sameFFN = no: -1.314— — (6)MFNdist ¡ 0.728: -0.718— — (6)MFNdist ≥ 0.728: 1.528— — (8)FFNdist ¡ 0.471: -0.863— — (8)FFNdist ≥ 0.471: -0.247— (1)sameFFN 6= no: 0.539— (2)sameFN = no: -1.475— — (7)FNdist ¡ 0.728: -0.371— — (7)FNdist ≥ 0.728: 1.241— (2)sameFN 6= no: 0.791— — (5)FFNdist ¡ 0.609: -0.734— — (5)FFNdist ≥ 0.609: 0.757— — (10)SNdist ¡ 0.738: -0.982— — (10)SNdist ≥ 0.738: 0.786— (3)B3dist ¡ 1.5: 1.142— (3)B3dist ≥ 1.5: -0.29— (4)LNdist ¡ 0.671: -1.454— (4)LNdist ≥ 0.671: -0.078— — (9)MNdist ¡ 0.606: -0.978— — (9)MNdist ≥ 0.606: 1.58

Table 7: Full dataset ADT model

23

Page 24: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

: -0.132— (1)sameFN = no: -1.461— — (9)FFNdist ¡ 0.539: -0.486— — (9)FFNdist ≥ 0.539: 0.433— (1)sameFN 6= no: 0.61— — (2)sameFFN = no: -1.296— — (2)sameFFN 6= no: 1.171— — (5)MFNDist ¡ 0.708: -0.962— — (5)MFNDist ≥ 0.708: 0.557— — (7)DPGeoDist ¡ 223: -0.532— — (7)DPGeoDist ≥ 223: 1.088— (3)B3dist ¡ 4.5: 0.674— — (10)B3dist ¡ 1.5: 0.431— — (10)B3dist ≥ 1.5: -0.379— (3)B3dist ≥ 4.5: -0.906— (4)LNdist ¡ 0.671: -1.292— (4)LNdist ≥ 0.671: -0.1— — (6)MNdist ¡ 0.606: -0.938— — (6)MNdist ≥ 0.606: 1.516— (8)SNdist ¡ 0.663: -0.765— (8)SNdist ≥ 0.663: 0.978

Table 8: ADT model without MV records

6.5. System Quality - Italian Dataset

In this section we examine the quality of results using the Italian dataset withdifferent configuration parameters. The configurable options were as follows.580

• Neighborhood Growth (NG): A parameter of the MFIBlocks algorithmcontrolling the amount of overlap allowed between clusters. The higherNG is, the more overlap may be present.

• MaxMinSup: As explained in Section 4.1, the MaxMinSup parameterdetermines the starting support level required by the MFI mining phase.585

MFIBlocks runs iteratively with decreasing minsup levels. For example,given MaxMinSup = 5, MFIBlocks would iteratively run with minsupvalues of {5,4,3,2}.

• Expert Weighting: Block score function can be weighted by item type.When this parameter is set to true, we use an expert derived weighting590

scheme rather than uniform item weights.• Expert Item Similarity (ExpertSim): In [18], block scores are cal-

culated based upon the Jaro-Winkler similarity function among items.While this method makes sense for q-grams derived from textual fields,we experiment with an expert knowledge-based function (Eq. 1), taking595

into account date part distance and geographical distance.• Same Source Discard (SameSrc): following the blocking stage, can-

didate pairs can be discarded if emanating from the same source. It isdeemed unlikely that the same person would appear twice in the samesource, since this implies that a person was named twice in the same vic-600

tim list or that a single witness filed two pages of testimony about thesame person.

• Classification (Cls): A binary parameter that when set to true the ADT

24

Page 25: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

classifier filters low scoring matches, rather than just output a similarityscore.605

A Cartesian product of all parameter values is prohibitively large. We there-fore report on the effect of running the full range of NG and MaxMinSup andthen fix NG and MaxMinSup to the values returning the best results and varyeach of the other conditions one at a time.

Figure 13: Meaningful False Positives - The Capelluto Children

Figure 14: Capelluto Family

Before delving into the results, we return610

to the motivation of this dataset, which is toenable automated construction of narrativesfrom multiple disparate sources. Figure 13captures two candidate pairs from the taggingapplication given to Yad Vashem researchers615

with candidate pairs from the MFIBlocks al-gorithm’s results. Elsa Capelluto, age 11 atdeath, is suggested to be matched with Giu-lia Capelluto (age 13), and Alberto Capel-luto (age 5). The records share a last name, fa-620

ther name, and mother name. Also (not shownin the image) all records are associated withthe island of Rhodes (Controlled by Italy atthe time, now part of Greece). While thesetwo candidate pairs are obvious false positives625

with respect to a single person entity matchingtask, they may be significant to users of YadVashem data seeking to create narratives forfamilies. Families often share many common properties, as the Capelluto fam-ily (Mother Zimbul and Sisters Elsa and Guilia pictured in Figure 14) demon-630

strates. Thus, we may wish to retain these pairs in a familial entity resolutiontask. These three pages of testimonies share a source, the aunt of these children,and thus they are discarded if the sameSrc feature is used.

Figure 15 presents the F-1 score of the different NG and MaxMinSup con-figurations. The horizontal axis designates the NG value used while each line635

corresponds to a different MaxMinSup value as specified by the legend. Resultsindicates peak F-1 score is obtained at an NG value of 3.5 for MaxMinSup = 4and a value of 3 for MaxMinSup = 5 and MaxMinSup = 6. However, whentaking into account that these results are to be filtered using a classifier and

25

Page 26: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

0.0

0.1

0.2

0.3

0.4

0.5

1.5 2 2.5 3 3.5 4 4.5 5

F-1

NG

MaxMinSup 6

MaxMinSup 5

MaxMinSup 4

Figure 15: F-1 score By NG and MaxMinSup

0.0

0.2

0.4

0.6

0.8

1.0

1.5 2 2.5 3 3.5 4 4.5 5

Pre

cisi

on

/ R

ecal

l

NG

Recall 6

Recall 5

Recall 4

Precision 6

Precision 5

Precision 4

Figure 16: Precision and Recall By NG and MaxMinSup

sameSrc, one would prefer higher Recall over Precision, and thus, an in-depth640

analysis (see Figure 16) reveals that in order to maximize Recall and with arelatively low price to pay in Precision, MaxMinSup = 5, and NG betweenthree and four are the preferred parameter settings. These values are in linewith previous findings [18].

We now report on results for the average over three runs where minsup ∈645

{5, 4, 3, 2} (MaxMinSup = 5) and NG varies between three and four (3, 3.5,4). Table 9 summarizes the experiments on the binary conditions defined above.

For all rows, the number reported indicate the average between three runswith varying NG. We provide the results for the null condition in the firstrow (Base) for reference. Applying expert weighting to item types provided a650

significant boost in Recall with only a minor drop in Precision and since otherconditions were expected to impact Recall negatively, we opted to continue therest of the experiments with this condition set to true. Using the hand craftedexpert similarity function was detrimental to both Precision and Recall. Usingthe two filter conditions (SameSrc and Cls) improved Precision at the expense655

of Recall, increasing the overall F-1 score up to 0.43 when both were used.

26

Page 27: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Condition Recall Precision F-1

Base 0.770 0.172 0.279Expert Weighting 0.886 0.151 0.256ExpertSim 0.872 0.118 0.205SameSrc 0.691 0.241 0.350Cls 0.850 0.204 0.328SameSrc + Cls 0.660 0.325 0.427

Table 9: Quality under Varying Conditions

Of the results reportedabove, only the negative ef-fect of using a hand craftedsimilarity function is sur-660

prising at first. However,one should keep in mindthat the suitability of theMFIBlocks algorithm to theentity resolution task hangs665

upon the special propertiesof the block scoring functionused, and specifically set-monotonicity [18]. This property is lost once a non-monotonic custom similarity function is used.

Finally, to better understand the algorithm’s misclassification, we manually670

evaluated 100 pairs from the set of about 1700 false positives. Of the pairsevaluated, we found 94 to be in fact true positives that were missing from thegolden standard. We therefore returned the full false positive set for taggingby Yad Vashem experts and expect the performance figures to be higher thanreported here in absolute terms while retaining the relative performance of the675

different conditions.

6.6. Comparative Quality of Blocking Algorithms

To empirically validate some of the arguments put forth in Section 4.1, wecompared the quality of MFIBlocks in terms of Precision and Recall with otherblocking algorithms. We base our comparison method on the one employed by680

Papadakis et. al. in their recent survey [24] and use their implementation ofthe comparison algorithms. For completeness, we now provide a short sum-mary of the definitions and succinct algorithm descriptions. For details, pleaserefer to the original work. Blocking techniques can be classified into three cat-egories, namely block building, which create blocks of records, block cleaning,685

which prune whole blocks, and comparison cleaning, which remove records fromblocks. In this work, we perform comparison cleaning through a highly specificclassification method by training an ADT model. To avoid giving our algorithman unfair advantage by way of this comparison method, we perform compari-son without classification. Furthermore, we use the default configuration of all690

methods, as suggested in [24].Table 10 presents the results of the comparative analysis. StBl stands for

the Standard Blocking [9, 23] technique, which creates a block for each attributevalue shared by more than one record. ACl refers to Attribute Clustering [23]which add a preliminary step to StBl in which similar tokens (e.g., John and695

Jhon) are grouped together by some similarity measure. Extended Sorted Neigh-borhood (ESoNe) [9] sorts the attribute values in alphabetical order and thenuses a sliding window of fixed size to create a block from all records which haveone of the values on the window. QGBl [15] is the Q-grams blocking techniquewhich adds a step to StBl where each attribute value is converted to all sub-700

sequences of q characters (q-grams). Extended Q-grams Blocking (EQBl) [9]

27

Page 28: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Blocking Algorithm Recall Precision

MFIBlocks 0.770 0.172StBl 1.0 < 0.001ACL 1.0 < 0.001CaCL 0.89 < 0.001ECaCL 0.876 0.003QGBl 0.998 < 0.001EQGBl 0.996 < 0.001ESoNe 1.0 < 0.001SuAr 0.796 0.003ESuAr 0.708 0.003TYPiMatch 0.717 < 0.001

Table 10: Comparative analysis of Blocking Techniques on Italy dataset

concatenates q-grams in an effort to increase the blocking keys’ discriminativeabilities. CaCl stands for Canpoy Clustering [21], a technique in which a ran-dom seed record is iteratively removed from a candidate pool and used to createa block using records which share the seed record’s attribute values. The keys705

for this block building methods are given by the QGBl method and it cre-ates non-overlapping blocks inherently by using non-replacing selection fromthe pool. ECaCl [9] extends CaCl by adding unassigned records to existingblocks. SuAr (Suffix Arrays) [1] attempts to improve StBl’s robustness tonoise by converting the attribute values to their maximal suffix of length larger710

than l. The extended version (ESuAr) [9] adds all of the attribute value’s sub-strings larger than l to the possible blocking keys. Finally, TYPiMatch [20]constructs a co-occurrence graph for all tokens and the maximal cliques are ex-tracted from it to create large blocks that are decomposed to smaller blocks bystandard blocking.715

In terms of precision, MFIBlocks dominates with a large margin of two ordersof magnitude over all other techniques, many of which give very imprecise results(less than 0.001). For some of the techniques, such as QGBl and CaCL, YadVashem data is not suitable, being pre-cleaned and with very few issues ofspelling mistakes. For others, we note that most of the meaningful attribute720

values (e.g., first-name code, gender-code, birth-day) are not names or wordsand thus cannot benefit from the similarity metrics and comparison methodsemployed by most compared techniques.

The best performance in recall terms is shared by StBl, ACL, and ESoNe.MFIBlocks performance is comparable with SUAr with 77% recall. It is worth725

noting that most of the blocking algorithms were designed for high recall, underthe assumption that blocking is merely a preprocessing phase to the ER process.This work, however, positions MFIBlocks as a tool for uncertain entity resolution,which calls for a more balanced precision/recall tradeoff, as is showcased inTable 10.730

28

Page 29: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

7. Conclusions and Open Questions

We presented a model for uncertain entity resolution, demonstrating it via aproject at Yad Vashem, which gave us a unique opportunity to apply state-of-the-art research prototypes to a real-life dataset. We have shown that MFIBlocksscales well with the size and complexity of this dataset. Quality of results for735

the Italy subset is encouraging and we hope foretells good results on the generaldataset as well. We have investigated alternatives in handling a Maybe tagand investigated a few methods for incorporating domain knowledge to improveupon the performance of the employed tools.

The model of uncertain entity resolution can generalize to many more do-740

mains beyond multi-sourced historical data. Nevertheless, by providing a deepanalysis of this model in a specific domain carries the benefits of highlightingunique characteristics in a way that is more accessible to researchers and indus-trial partners alike. We believe that the proposed uncertain entity resolutionmodel carries much potential in the new realms of big data, where volume,745

variety, and veracity of data pose new challenges to the database community.The question of potential benefits and appropriate methods of incorporating

domain knowledge in entity resolution is far from resolved. Furthermore, theunique properties of this project prompted us to pose some additional questionswhich were only partially addressed within the limited scope of this paper. How750

can we exploit implicit and explicit knowledge about record sources in the multi-source setting? Can we effectively perform entity resolution on different levelsof resolution, e.g., families in this dataset. The inter-record relationships in thisdataset open the door to questions such as how to perform entity resolution atthe edge and sub-graph level and not just at the node level? We leave these755

questions for future research.For Yad Vashem, this project is a corner-stone for future applications that

will enable researchers and the general public to gain insight into the livesand tragedies of Holocaust-era individuals and communities. For example, YadVashem has sponsored a Hackathon4 in which many of the submitted applica-760

tions were based upon the ability to resolve entities and suggested methods toautomatically extract narratives from Yad Vashem data. At the time of publi-cation, Yad Vashem is actively engaged in integrating the results of the projectinto its databases and applications.

Acknowledgments765

We wish to extend our thanks to Sapir Golan for his invaluable technicalassistance.

[1] A. Aizawa and K. Oyama. A fast linkage detection scheme for multi-sourceinformation integration. In WIRI ’05: Proceedings of the International

4https://youtu.be/sgoVF5qKuPQ Innovation in the Service of Memory, retrieved February17th, 2016

29

Page 30: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

Workshop on Challenges in Web Information Retrieval and Integration,770

pages 30–39, Washington, DC, USA, 2005. IEEE Computer Society.

[2] P. Andritsos, A. Fuxman, and R. Miller. Clean answers over dirtydatabases: A probabilistic approach. In ICDE, page 30, 2006.

[3] G. Beskales, M. A. Soliman, I. F. Ilyas, and S. Ben-David. Modelingand querying possible repairs in duplicate detection. Proc. VLDB Endow.,775

2(1):598–609, 2009.

[4] M. Bilenko and R. J. Mooney. Adaptive duplicate detection using learnablestring similarity measures. In Workshop on Data Cleaning, Record Linkage,and Object Consolidation. ACM SIGKDD, 2003.

[5] T. Blanke and C. Kristel. Integrating Holocaust Research. International780

Journal of Humanities and Arts Computing, 7(1-2):41–57, 2013.

[6] C. Borgelt. An implementation of the fp-growth algorithm. In Proceedingsof the 1st international workshop on open source data mining: frequentpattern mining implementations, pages 1–5. ACM, 2005.

[7] S. Chaudhuri, V. Ganti, and R. Motwani. Robust identification of fuzzy785

duplicates. In ICDE05, pages 865–876, 2005.

[8] P. Christen. FEBRL: a freely available record linkage system with a graph-ical user interface. In HDKM, 2008.

[9] P. Christen. A Survey of Indexing Techniques for Scalable Record Linkageand Deduplication. IEEE Transactions on Knowledge and Data Engineer-790

ing, 24(9):1537–1555, 2012.

[10] W. W. Cohen. Data integration using similarity joins and a word-basedinformation representation language. ACM Trans. Inf. Syst., 18:288–321,July 2000.

[11] A. K. Elmagarmid, P. G. Ipeirotis, and V. S. Verykios. Duplicate record795

detection: A survey. IEEE Transactions on Knowledge and Data Engineer-ing, 19(1):1–16, 2007.

[12] I. P. Fellegi and A. B. Sunter. A theory for record linkage. Journal of theAmerican Statistical Association, 64(328), 1969.

[13] Freund, Yoav and L. Mason. The alternating decision tree learning algo-800

rithm. International Conference on Machine Learning, 99:124–133, 1999.

[14] A. Gal. Tutorial: Uncertain entity resolution. PVLDB, 7(13):1711–1712,2014.

[15] L. Gravano, P. G. Ipeirotis, H. V. Jagadish, N. Koudas, S. Muthukrishnan,D. Srivastava, et al. Approximate string joins in a database (almost) for805

free. In VLDB, volume 1, pages 491–500, 2001.

30

Page 31: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

[16] M. Hall, H. National, E. Frank, G. Holmes, B. Pfahringer, P. Reute-mann, and I. H. Witten. The WEKA Data Mining Software : An Update.SIGKDD Explorations, 11(1):10–18, 2009.

[17] E. Ioannou, W. Nejdl, C. Niederee, and Y. Velegrakis. On-the-fly entity-810

aware query processing in the presence of linkage. PVLDB, 3(1):429–438,2010.

[18] B. Kenig and A. Gal. MFIBlocks: An effective blocking algorithm for entityresolution. Information Systems, 38(6):908–926, Sept. 2013.

[19] H. Kopcke, A. Thor, and E. Rahm. Evaluation of entity resolution ap-815

proaches on real-world match problems. Proceedings of the VLDB Endow-ment, 3(1):484–493, 2010.

[20] Y. Ma and T. Tran. Typimatch: Type-specific unsupervised learning ofkeys and key values for heterogeneous web data integration. In Proceedingsof the sixth ACM international conference on Web search and data mining,820

pages 325–334. ACM, 2013.

[21] A. McCallum, K. Nigam, and L. H. Ungar. Efficient clustering of high-dimensional data sets with application to reference matching. In KDD’00: Proceedings of the sixth ACM SIGKDD international conference onKnowledge discovery and data mining, pages 169–178, New York, NY, USA,825

2000. ACM.

[22] F. Naumann and M. Herschel. An introduction to duplicate detection.Synthesis Lectures on Data Management, 2(1):1–87, 2010.

[23] G. Papadakis, E. Ioannou, T. Palpanas, C. Niederee, and W. Nejdl. Ablocking framework for entity resolution in highly heterogeneous infor-830

mation spaces. IEEE Transactions on Knowledge and Data Engineering,25(12):2665–2682, 2013.

[24] G. Papadakis, J. Svirsky, A. Gal, and T. Palpanas. Comparative analysis ofapproximate blocking techniques for entity resolution. Proc. VLDB Endow.,9(9):684–695, May 2016.835

[25] S. Rong, X. Niu, E. W. Xiang, H. Wang, Q. Yang, and Y. Yu. A machinelearning approach for instance matching based on similarity metrics. InThe Semantic Web - ISWC 2012 - 11th International Semantic Web Con-ference, Boston, MA, USA, November 11-15, 2012, Proceedings, Part I,pages 460–475, 2012.840

[26] S. Sarawagi and A. Bhamidipaty. Interactive deduplication using activelearning. In KDD, pages 269–278, 2002.

[27] W. Winkler. Methods for record linkage and bayesian networks. TechnicalReport RRS2002/05, US Bureau of the Census, Washington, D.C., 2002.Statistical Research Report Series.845

31

Page 32: Multi-Source Uncertain Entity Resolution: Transforming Holocaust … · 2016-12-22 · Multi-Source Uncertain Entity Resolution: Transforming Holocaust Victim Reports into People

[28] W. E. Winkler. Overview of record linkage and current research directions.Current, (2006-2):1–28, 2006.

[29] C. Xiao, W. Wang, X. Lin, and J. Yu. Similarity joins for near duplicatedetection. In WWW, 2008.

32