how well does your instance matching system perform? experimental evaluation with lance

19
How well does your Instance Matching system perform? Experimental evaluation with LANCE Tzanina Saveta, Evangelia Daskalaki, Giorgos Flouris, Irini Fundulaki Institute of Computer Science – FORTH, Greece Axel-Cyrille Ngonga Ngomo IFI/AKSW, University of Leipzig, Germany 10/31/16 ISWC 2016: How well does your Instance Matching system perform? Experimental evaluation with LANCE 1

Upload: holistic-benchmarking-of-big-linked-data

Post on 14-Apr-2017

231 views

Category:

Science


0 download

TRANSCRIPT

Page 1: How well does your Instance Matching system perform? Experimental evaluation with LANCE

HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE

TzaninaSaveta,EvangeliaDaskalaki,GiorgosFlouris,

IriniFundulakiInstituteofComputerScience–FORTH,Greece

Axel-CyrilleNgongaNgomoIFI/AKSW,UniversityofLeipzig,Germany

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 1

Page 2: How well does your Instance Matching system perform? Experimental evaluation with LANCE

WhyInstanceMatching?

ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 2*AdaptedfromSuchanek&Weikumtutorial@SIGMOD2013

Differentsourcescontaindifferentdescriptionsofthesamerealworld

entity

Page 3: How well does your Instance Matching system perform? Experimental evaluation with LANCE

InstanceMatchingforLinkedData

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 3

SetofRDFtriplesconstituteanRDF

graph

SparseData

Richsemanticsexpressedinterms

ofontologies

LargenumberofsourcestointegrateValue,Structure

andSemanticsHeterogeneities

*AdaptedfromSuchanek&Weikumtutorial@SIGMOD2013

Page 4: How well does your Instance Matching system perform? Experimental evaluation with LANCE

Benchmarking

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 4

Instancematchinghasledtothedevelopmentofanumberofmatchingtechniquesandtools

•  Howtocomparethose?•  Howtoassesstheirperformance(efficiencyand

effectiveness)?•  Howto“push”systemsintobecomingbetter?

•  Benchmarkyoursystems!

Page 5: How well does your Instance Matching system perform? Experimental evaluation with LANCE

InstanceMatchingBenchmarkComponents

•  Datasets–  Sourceandthetargetdatasetsthatwillbematchedtogethertofindtheentitiesthatrefertothesamerealworldobject

•  Groundtruth/Goldstandard/Referencealignment–  The“correctanswersheet”usedtojudgethecompletenessandsoundnessoftheresultsproducedbytheSUT

•  Organizedintotestcaseseachaddressingdifferentkindofinstancematchingrequirements

•  Metrics–  Theperformancemetric(s)thatdeterminethesystems’efficiencyandeffectiveness

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 5

Page 6: How well does your Instance Matching system perform? Experimental evaluation with LANCE

LANCE

•  Anovelinstancematchingbenchmarkgenerator

•  Domain-independent

•  Highlyconfigurableandscalable•  Standardvalue-basedandstructure-basedtestcases•  Advancedsemantics-awaretestcasesconsideringOWL2

expressiveconstructs

•  Richweightedgoldstandard

•  Additionalmetrics:similarityscoremetric

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 6

Page 7: How well does your Instance Matching system perform? Experimental evaluation with LANCE

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 7

LANCEArchitecture

Source Data

Target Data

Weighted Gold Standard

Resource Transformation

Module

RESCAL [NT12]

MATCHER SAMPLER

Weight Computation Module

Test Case Generation Parameters RDF

Repository Dat

a

Inge

stio

n M

odul

e

Initialization Module

Resource Generator

Test Case Generator SP

ARQ

L Q

uerie

s (S

chem

a St

ats)

SPAR

QL

Que

ries

(IR)

Matched Instances

Source Data

Page 8: How well does your Instance Matching system perform? Experimental evaluation with LANCE

TestCases

Testcasesarebuiltusingavarietyoftransformations

•  Value-basedtestcases–  Transformationsofvaluesofdatatypeproperties

•  Structure-basedtestcases–  Transformationsofstructureofobjectanddatatypeproperties

•  Semantics-awaretestcases–  Transformationsattheinstancelevelconsideringtheschema

•  SimpleandComplexcombinationofthethreefirstcategories

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 8

Page 9: How well does your Instance Matching system perform? Experimental evaluation with LANCE

LANCEPerformanceMetrics•  Averagesimilarityscore:averagedifficultyofthematchedinstances

–  Benchmarkwithhighaveragesimilarityscore:matchedinstancesareeasiertofind

•  Standarddeviation:spreadofsimilarityscoresforthematchedinstances–  Benchmarkwithhighstandarddeviation:

•  scoresarespreadoutfromtheaverage•  moreheterogeneityofmatchedinstances

10/31/16 HOBBITPlenary2

Obtainamorefine-grainedunderstandingoftheIMsystem’sperformancebycomparingtheaveragestandarddeviationand

similarityscoreofthesystemandbenchmark

Page 10: How well does your Instance Matching system perform? Experimental evaluation with LANCE

Experiments•  EfficiencyandeffectivenessofIMsystemsusingLANCEbenchmarks–  Systems:•  LogMapVersion2.4[JG11](MoReReasoner[RG13])•  OtO[DP12]•  LIMES(EAGLEIMalgorithm[NL12])

–  Datasets•  LDBC’sSPIMBENCHGenerator(SemanticPublishingBenchmark)

•  UOBM– MatchingTask•  All5categoriesintroducedpreviously•  Allinstancesweretransformed

10

Page 11: How well does your Instance Matching system perform? Experimental evaluation with LANCE

SPIMBENCH:StandardMetrics

11

•  LogMap–  Respondwellinthevalue-basedtestcases–  Reducedperformancewhenalsosemantics-awaretestcaseswereapplied

Page 12: How well does your Instance Matching system perform? Experimental evaluation with LANCE

SPIMBENCH:StandardMetrics

12

•  OtOandEAGLE–  Givegoodresultsregardingthevalue-basedtransformations

–  Reducedperformanceintheremainingcategories•  EAGLEisnon-deterministicandusesunsupervisedlearning

Page 13: How well does your Instance Matching system perform? Experimental evaluation with LANCE

UOBM:StandardMetrics

•  LogMap1.Doesnotperformwelltoanyofthecategories2.Performancenotaffectedbythedatasetsize•  OtO1.Performsbetter2.Reducedperformancewhenincreasingdatasetsize

13

Page 14: How well does your Instance Matching system perform? Experimental evaluation with LANCE

SPIMBENCH:AdditionalMetrics

DistributionofsimilarityscoresforLANCEandTruePositivematchesfromIMsystemsforsemantics-awaretestcasesinthecaseofthe10Ktriplesdataset.•  LogMapcanaddressdifficulttestcases•  EAGLE&OtOcanaddressmostlyvalue-basedtestcases

1

10

100

0.7 0.72 0.74 0.76 0.78 0.8 0.82 0.84 0.86 0.88 0.9 0.92 0.94 0.96 0.98 1

log(#ofm

appings)

SimilarityScores

OtO EAGLE LogMap LANCE

14

StandardDevia8on

Page 15: How well does your Instance Matching system perform? Experimental evaluation with LANCE

UOBM:AdditionalMetrics

DistributionofsimilarityscoresforLANCEandTruePositivematchesfromIMsystemsforstructure-basedtestcasesinthecaseofthe10Ktriplesdataset.•  LogMapcannotaddresswellthechangeofURIsintheInstances

ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 15

1

10

100

0.6 0.62 0.64 0.66 0.68 0.7 0.72 0.74 0.76 0.78 0.8 0.82 0.84 0.86 0.88 0.9

log(#ofm

appings)

SimilarityOtO LogMap LANCE

0

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

OtO LogMap LANCE

Page 16: How well does your Instance Matching system perform? Experimental evaluation with LANCE

LessonsLearned•  DifferenttypeoftransformationsaffectIMsystem’s

performance•  Thecharacteristicsofsourcedatasetsaffectthebehaviorof

IMsystems

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 16

Page 17: How well does your Instance Matching system perform? Experimental evaluation with LANCE

Questions?

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 17

Page 18: How well does your Instance Matching system perform? Experimental evaluation with LANCE

AcknowledgmentsThisprojecthasreceivedfundingfromtheEuropeanUnion’sHorizon2020researchandinnovationprogrammeundergrantagreementNo688227.

10/31/16 ISWC2016:HowwelldoesyourInstanceMatchingsystemperform?ExperimentalevaluationwithLANCE 18

Page 19: How well does your Instance Matching system perform? Experimental evaluation with LANCE

References[JG11]E.Jimenez-RuizandB.C.Grau.Logmap:Logic-basedandscalableontologymatching.InISWC,2011.[RG13]A.A.Romero,B.C.Grau,etal.MORe:aModularOWLReasonerforOntologyClassification.InORE,pages61-67,2013.[DP12]E.DaskalakiandD.Plexousakis.OtOMatchingSystem:AMulti-strategyApproachtoInstanceMatching.InCAiSE,2012.[NL12]A.-C.NgongaNgomoandK.Lyko.EAGLE:EfficientActiveLearningofLinkSpecificationsusingGeneticProgramming.InESWC,2012.

19