SlideShare uma empresa Scribd logo
1 de 23
Baixar para ler offline
Don’t compare Apples to Oranges -
Extending GERBIL for a fine grained NEL Evaluation
Jörg Waitelonis, Henrik Jürges, Harald Sack
Hasso-Plattner-Institute for IT-Systems Engineering, University of Potsdam
Semantics 2016, Leipzig, Germany, September 12-15th, 2016
Agenda
1. NEL and NEL evaluation
2. Dataset properties and evaluation drawbacks
3. Extending GERBIL
● Building conditional datasets
● Measure dataset characteristics
1. Results
2. Demonstration
3. Summary & Future work
Named Entity Linking (NEL)
Chart 3
… Armstrong …
Named Entity Linking (NEL), Principle
Chart 4
“Armstrong landed on the moon.”
Candidates:
dbr:Neil_Armstrong
dbr:Lance_Armstrong
dbr:Louis_Armstrong
….
Candidates:
dbr:Moon
dbr:Lunar
….
Correct entities
Entity mention with
surface form
String Distance
Link Analysis
Vector Space
Fuzzy String Matching
Conditional Random Fields
Random Forest
RankSVM
Learning to Rank
Surface Aggregation
Word Embeddings
Context Similarity Matching
1. Tokenize text
2. Find candidates in KB
3. Score candidates with a
magic algorithm and
select the best one
KEA
Wikifier
● Algorithm only approximates correct
entities
● Need for verification and testing
● Dataset consists out of:
■ Documents (String/Sentences)
■ Annotations (ground truth)
Named Entity Linking, Evaluation
Chart 5
ACE2004
AIDA/CoNLL
DBpedia Spotlight
IITB
KORE50
MSNBC
Micropost2014
N3-RSS-500
WES2015
N3-Reuters-128
● Traditional measures are:
■ Precision: defines how well an annotator works
■ Recall: defines how complete the results are
■ F1-measure: harmonic medium between precision and recall
■ And more, cf. Rizzo et al. [1]
● GERBIL - a general entity annotation system (AKSW Leipzig), cf.
Usbeck et. al [2]
● Used for testing/optimizing/benchmarking annotators
● Neat Webinterface
● 13 Annotators / 20 datasets
● F-measure to rough for detailed
evaluation
● Developer need dataset insights
Named Entity Linking, Benchmarking
Chart 6
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
● Size of a datasets
● Amount of annotations/documents/words
● What types of entities are used? E.g. persons, places, events, ….
● Are there documents without annotations? E.g. Microposts 2014
● What sort of popularity have the entities? E.g. PageRank, Indegree
● How ambiguous are the entities and surface forms?
● How divers are the entities and surface forms?
● ….
Cf. van Erp et al. [3]
Properties of Datasets
Chart 7
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
● How does the dataset characteristics influence the evaluation
results?
● How does the popularity of entities influence the evaluation
results?
● How can a general dataset be used for domain specific NEL tools?
● How can datasets be compared? Is there something like a general
difficulty?
● Limited comparability between benchmark results
● Penalization of good annotators with inappropriate datasets
Cf. van Erp et al. [3]
Research Questions and Drawbacks
Chart 8
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
● Approach for a solution:
● Adjustable filter system for GERBIL
● Expose dataset characteristics
● Datasets and annotators added at runtime are also included
● Visualize the results
Extending GERBIL
Chart 9
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
Extending GERBIL, Conditional Datasets
Chart 10
Dataset
Type and popularity
specific datasets
Annotator
Documents Annotator results
Benchmark results
Evaluate each specific
dataset and result
PR(e) > t
PR(e) > t
PR(e) > t
rdf:type
rdf:type
rdf:type
rdf:type
rdf:type
rdf:type
Results, Types
Chart 11
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
Results, Popularity
Chart 12
Extending GERBIL, Not Annotated Documents
● Not annotated documents: shows the relative amount of empty
documents within a datasets
● Only affects if annotators searches entity mentions by themselves
Chart 13
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
Extending GERBIL, Density
Chart 14
● Density: shows the relation between number of annotations and
words in the document
● Only affects if annotators searches entity mentions by themselves
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
Extending GERBIL, Likelihood of Confusion
● Likelihood of Confusion (Level of Ambiguity)
● True measures are unknown due to missing exhaustive collections
● Rough overview how difficult to disambiguate
Chart 15
Entities Surface Forms
Tegel
TXL
Bruce
Otto Lilienthal
Bruce Lee
Bruce Willis
Airport Tegel
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
synonyms
Results, Likelihood of Confusion
Chart 16
Entities
● A high red bar indicates
an entity has a high
amount of homonyms
● A high blue bar indicates
a surface form has a high
amount of synonyms
Surface Forms
Extending GERBIL, Dominance of Entities
Chart 17
Bruce
Bruci
Bruce Willis
Testdata
Vocabulary
dominance(e)=
e(t)/e(v)
● Expresses the relation between
used words and all words
● True measures unknown
● High rates prevents overfitting
● Prevents repetition of surface
forms
dbr:Bruce_Willis
Bruce Walter Willis
Extending GERBIL, Dominance of Surface Forms
Chart 18Chart 18
dbr:Irene_Angelina
dbr:Angelina_Jordan
dbr:Angelina_Jolie
Vocabulary
dominance(s)=
s(t)/s(v)
● Expresses the relation between
used mentions and all
mentions
● True measures unknown
● High rates prevents
overfitting
● Indicates how context
dependent a disambiguation is
Testdata
Angelina
Results, Dominance
Chart 19
● Blue bar indicates that for a
entity a variety of surface
forms is used
● Red bar indicates how context
dependent the disambiguation of
an surface form is
Dominance of surface forms Dominance of entities
Demo
● http://gerbil.s16a.org/
● https://github.com/santifa/gerbil/
Chart 20
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
■ Summary:
□ Implemented a domain specific filter system
□ Measure dataset characteristics
□ Annotator results are nearly the same on entities of different
popularity
□ Enable specific analyses and optimization of annotators
□ Enable users to select the tools that performs best for a specific
domain
■ Future work:
□ Keep up with GERBIL development, increase performance
□ More measurements, e. g. max_recall
□ Dataset remixing ≙ assemble new customized datasets
– E. g. Unpopular companies
Summary & Future work
Chart 21
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
[1] Giuseppe Rizzo, Amparo Elizabeth Cano Basave, Bianca Pereira, and Andrea
Varga. Making Sense of Microposts (#Microposts2015) Named Entity rEcognition and
Linking (NEEL) Challenge. In 5th Workshop on Making Sense of Microposts
(#Microposts2015), pages 44–53. CEUR-WS.org, 2015
[2] M. Röder, R. Usbeck, and A.-C. Ngonga Ngomo. Gerbil’s new stunts: Semantic
annotation benchmarking improved. Technical report, Leipzig University, 2016
[3] M. van Erp, P. Mendes, H. Paulheim, F. Ilievski, J. Plu, G. Rizzo, and J. Waitelonis.
Evaluating entity linking: An analysis of current benchmark datasets and a roadmap
for doing a better job. In Proc. of 10th edition of the Language Resources and
Evaluation Conference, Portoroz, Slovenia, 2016.
References
Chart 22
Henrik Jürges
Semantics 2016
Leipzig, Germany
Don’t compare
Apples to Oranges:
Extending GERBIL
for a fine grained
NEL Evaluation
Questions?
Questions?
Thank you for your attention!

Mais conteúdo relacionado

Destaque

Michael Fuchs | How to compute semantic relationships between entities and fa...
Michael Fuchs | How to compute semantic relationships between entities and fa...Michael Fuchs | How to compute semantic relationships between entities and fa...
Michael Fuchs | How to compute semantic relationships between entities and fa...semanticsconference
 
Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...
Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...
Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...semanticsconference
 
Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...
Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...
Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...semanticsconference
 
Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...
Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...
Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...semanticsconference
 
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...semanticsconference
 
Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...
Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...
Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...semanticsconference
 
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...semanticsconference
 
Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...
Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...
Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...semanticsconference
 
Kostas Kastrantas | Business Opportunities with Linked Open Data
Kostas Kastrantas  | Business Opportunities with Linked Open DataKostas Kastrantas  | Business Opportunities with Linked Open Data
Kostas Kastrantas | Business Opportunities with Linked Open Datasemanticsconference
 
David Kuilman | Creating a Semantic Enterprise Content model to support conti...
David Kuilman | Creating a Semantic Enterprise Content model to support conti...David Kuilman | Creating a Semantic Enterprise Content model to support conti...
David Kuilman | Creating a Semantic Enterprise Content model to support conti...semanticsconference
 
Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...
Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...
Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...semanticsconference
 
Victor Charpenay | Standardized Semantics for an Open Web of Things
Victor Charpenay | Standardized Semantics for an Open Web of ThingsVictor Charpenay | Standardized Semantics for an Open Web of Things
Victor Charpenay | Standardized Semantics for an Open Web of Thingssemanticsconference
 
Chalitha Perera | Cross Media Concept and Entity Driven Search for Enterprise
Chalitha Perera | Cross Media Concept and Entity Driven Search for EnterpriseChalitha Perera | Cross Media Concept and Entity Driven Search for Enterprise
Chalitha Perera | Cross Media Concept and Entity Driven Search for Enterprisesemanticsconference
 
Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...
Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...
Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...semanticsconference
 
Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...
Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...
Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...semanticsconference
 
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...semanticsconference
 
Thomas Vavra | New Ways of Handling Old Data
Thomas Vavra | New Ways of Handling Old DataThomas Vavra | New Ways of Handling Old Data
Thomas Vavra | New Ways of Handling Old Datasemanticsconference
 
OOPS!: on-line ontology diagnosis by Maria Poveda
OOPS!: on-line ontology diagnosis by Maria PovedaOOPS!: on-line ontology diagnosis by Maria Poveda
OOPS!: on-line ontology diagnosis by Maria Povedasemanticsconference
 
Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...
Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...
Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...semanticsconference
 
Felix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINE
Felix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINEFelix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINE
Felix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINEsemanticsconference
 

Destaque (20)

Michael Fuchs | How to compute semantic relationships between entities and fa...
Michael Fuchs | How to compute semantic relationships between entities and fa...Michael Fuchs | How to compute semantic relationships between entities and fa...
Michael Fuchs | How to compute semantic relationships between entities and fa...
 
Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...
Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...
Joe Pairman | Multiplying the Power of Taxonomy with Granular, Structured Con...
 
Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...
Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...
Sebastian Bader | Semantic Technologies for Assisted Decision-Making in Indus...
 
Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...
Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...
Camilo Thorne, Stefano Faralli and Heiner Stuckenschmidt | Entity Linking for...
 
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...
 
Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...
Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...
Vladimir Alexiev | Semantic Enrichment of Twitter Microposts Helps Understand...
 
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
 
Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...
Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...
Najmeh Mousavi Nejad, Simon Scerri, Sören Auer and Elisa M. Sibarani | EULAid...
 
Kostas Kastrantas | Business Opportunities with Linked Open Data
Kostas Kastrantas  | Business Opportunities with Linked Open DataKostas Kastrantas  | Business Opportunities with Linked Open Data
Kostas Kastrantas | Business Opportunities with Linked Open Data
 
David Kuilman | Creating a Semantic Enterprise Content model to support conti...
David Kuilman | Creating a Semantic Enterprise Content model to support conti...David Kuilman | Creating a Semantic Enterprise Content model to support conti...
David Kuilman | Creating a Semantic Enterprise Content model to support conti...
 
Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...
Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...
Shuangyong Song, Qingliang Miao and Yao Meng | Linking Images to Semantic Kno...
 
Victor Charpenay | Standardized Semantics for an Open Web of Things
Victor Charpenay | Standardized Semantics for an Open Web of ThingsVictor Charpenay | Standardized Semantics for an Open Web of Things
Victor Charpenay | Standardized Semantics for an Open Web of Things
 
Chalitha Perera | Cross Media Concept and Entity Driven Search for Enterprise
Chalitha Perera | Cross Media Concept and Entity Driven Search for EnterpriseChalitha Perera | Cross Media Concept and Entity Driven Search for Enterprise
Chalitha Perera | Cross Media Concept and Entity Driven Search for Enterprise
 
Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...
Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...
Stephen Buxton | Data Integration - a Multi-Model Approach - Documents and Tr...
 
Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...
Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...
Kolawole John Adebayo, Luigi Di Caro and Guido Boella | A Supervised Keyphras...
 
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
 
Thomas Vavra | New Ways of Handling Old Data
Thomas Vavra | New Ways of Handling Old DataThomas Vavra | New Ways of Handling Old Data
Thomas Vavra | New Ways of Handling Old Data
 
OOPS!: on-line ontology diagnosis by Maria Poveda
OOPS!: on-line ontology diagnosis by Maria PovedaOOPS!: on-line ontology diagnosis by Maria Poveda
OOPS!: on-line ontology diagnosis by Maria Poveda
 
Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...
Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...
Georgios Meditskos and Stamatia Dasiopoulou | Question Answering over Pattern...
 
Felix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINE
Felix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINEFelix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINE
Felix Burkhardt | ARCHITECTURE FOR A QUESTION ANSWERING MACHINE
 

Semelhante a Jörg Waitelonis, Henrik Jürges and Harald Sack | Don't compare Apples to Oranges - Extending GERBIL for a fine grained NEL evaluation

Babak Rasolzadeh: The importance of entities
Babak Rasolzadeh: The importance of entitiesBabak Rasolzadeh: The importance of entities
Babak Rasolzadeh: The importance of entitiesZoltan Varju
 
TFMAP: Optimizing MAP for Top-N Context-aware Recommendation
TFMAP: Optimizing MAP for Top-N Context-aware RecommendationTFMAP: Optimizing MAP for Top-N Context-aware Recommendation
TFMAP: Optimizing MAP for Top-N Context-aware RecommendationAlexandros Karatzoglou
 
Improving Machine Learning using Graph Algorithms
Improving Machine Learning using Graph AlgorithmsImproving Machine Learning using Graph Algorithms
Improving Machine Learning using Graph AlgorithmsNeo4j
 
MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...
MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...
MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...Tao Xie
 
General introduction to AI ML DL DS
General introduction to AI ML DL DSGeneral introduction to AI ML DL DS
General introduction to AI ML DL DSRoopesh Kohad
 
The best stats you've ever seen
The best stats you've ever seenThe best stats you've ever seen
The best stats you've ever seenParul Verma
 
Improve ml predictions using graph algorithms (webinar july 23_19).pptx
Improve ml predictions using graph algorithms (webinar july 23_19).pptxImprove ml predictions using graph algorithms (webinar july 23_19).pptx
Improve ml predictions using graph algorithms (webinar july 23_19).pptxNeo4j
 
Machine_Learning_Project_Report
Machine_Learning_Project_ReportMachine_Learning_Project_Report
Machine_Learning_Project_ReportAditya Hendra
 
Employing Graph Databases as a Standardization Model towards Addressing Heter...
Employing Graph Databases as a Standardization Model towards Addressing Heter...Employing Graph Databases as a Standardization Model towards Addressing Heter...
Employing Graph Databases as a Standardization Model towards Addressing Heter...Dippy Aggarwal
 
Tableau Final Presentation
Tableau Final PresentationTableau Final Presentation
Tableau Final PresentationAnvesh Rao
 
ArXiv Literature Exploration using Social Network Analysis
ArXiv Literature Exploration using Social Network AnalysisArXiv Literature Exploration using Social Network Analysis
ArXiv Literature Exploration using Social Network AnalysisTanat Iempreedee
 
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...Codiax
 

Semelhante a Jörg Waitelonis, Henrik Jürges and Harald Sack | Don't compare Apples to Oranges - Extending GERBIL for a fine grained NEL evaluation (20)

Babak Rasolzadeh: The importance of entities
Babak Rasolzadeh: The importance of entitiesBabak Rasolzadeh: The importance of entities
Babak Rasolzadeh: The importance of entities
 
TFMAP: Optimizing MAP for Top-N Context-aware Recommendation
TFMAP: Optimizing MAP for Top-N Context-aware RecommendationTFMAP: Optimizing MAP for Top-N Context-aware Recommendation
TFMAP: Optimizing MAP for Top-N Context-aware Recommendation
 
Improving Machine Learning using Graph Algorithms
Improving Machine Learning using Graph AlgorithmsImproving Machine Learning using Graph Algorithms
Improving Machine Learning using Graph Algorithms
 
Learning to learn Model Behavior: How to use "human-in-the-loop" to explain d...
Learning to learn Model Behavior: How to use "human-in-the-loop" to explain d...Learning to learn Model Behavior: How to use "human-in-the-loop" to explain d...
Learning to learn Model Behavior: How to use "human-in-the-loop" to explain d...
 
Alan Berg
Alan Berg Alan Berg
Alan Berg
 
MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...
MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...
MSRA 2018: Intelligent Software Engineering: Synergy between AI and Software ...
 
General introduction to AI ML DL DS
General introduction to AI ML DL DSGeneral introduction to AI ML DL DS
General introduction to AI ML DL DS
 
L15.pptx
L15.pptxL15.pptx
L15.pptx
 
Icpc13.ppt
Icpc13.pptIcpc13.ppt
Icpc13.ppt
 
The best stats you've ever seen
The best stats you've ever seenThe best stats you've ever seen
The best stats you've ever seen
 
Improve ml predictions using graph algorithms (webinar july 23_19).pptx
Improve ml predictions using graph algorithms (webinar july 23_19).pptxImprove ml predictions using graph algorithms (webinar july 23_19).pptx
Improve ml predictions using graph algorithms (webinar july 23_19).pptx
 
Machine_Learning_Project_Report
Machine_Learning_Project_ReportMachine_Learning_Project_Report
Machine_Learning_Project_Report
 
Employing Graph Databases as a Standardization Model towards Addressing Heter...
Employing Graph Databases as a Standardization Model towards Addressing Heter...Employing Graph Databases as a Standardization Model towards Addressing Heter...
Employing Graph Databases as a Standardization Model towards Addressing Heter...
 
Tableau Final Presentation
Tableau Final PresentationTableau Final Presentation
Tableau Final Presentation
 
Tableau Presentation
Tableau PresentationTableau Presentation
Tableau Presentation
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component Analysis
 
ArXiv Literature Exploration using Social Network Analysis
ArXiv Literature Exploration using Social Network AnalysisArXiv Literature Exploration using Social Network Analysis
ArXiv Literature Exploration using Social Network Analysis
 
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
 
Sub1579
Sub1579Sub1579
Sub1579
 
Graph based data models
Graph based data modelsGraph based data models
Graph based data models
 

Mais de semanticsconference

Linear books to open world adventure
Linear books to open world adventureLinear books to open world adventure
Linear books to open world adventuresemanticsconference
 
Session 1.2 high-precision, context-free entity linking exploiting unambigu...
Session 1.2   high-precision, context-free entity linking exploiting unambigu...Session 1.2   high-precision, context-free entity linking exploiting unambigu...
Session 1.2 high-precision, context-free entity linking exploiting unambigu...semanticsconference
 
Session 4.3 semantic annotation for enhancing collaborative ideation
Session 4.3   semantic annotation for enhancing collaborative ideationSession 4.3   semantic annotation for enhancing collaborative ideation
Session 4.3 semantic annotation for enhancing collaborative ideationsemanticsconference
 
Session 1.1 dalicc - data licenses clearance center
Session 1.1   dalicc - data licenses clearance centerSession 1.1   dalicc - data licenses clearance center
Session 1.1 dalicc - data licenses clearance centersemanticsconference
 
Session 1.3 context information management across smart city knowledge domains
Session 1.3   context information management across smart city knowledge domainsSession 1.3   context information management across smart city knowledge domains
Session 1.3 context information management across smart city knowledge domainssemanticsconference
 
Session 0.0 aussenac semanticsnl-pwebsem2017-v4
Session 0.0   aussenac semanticsnl-pwebsem2017-v4Session 0.0   aussenac semanticsnl-pwebsem2017-v4
Session 0.0 aussenac semanticsnl-pwebsem2017-v4semanticsconference
 
Session 0.0 keynote sandeep sacheti - final hi res
Session 0.0   keynote sandeep sacheti - final hi resSession 0.0   keynote sandeep sacheti - final hi res
Session 0.0 keynote sandeep sacheti - final hi ressemanticsconference
 
Session 1.1 linked data applied: a field report from the netherlands
Session 1.1   linked data applied: a field report from the netherlandsSession 1.1   linked data applied: a field report from the netherlands
Session 1.1 linked data applied: a field report from the netherlandssemanticsconference
 
Session 1.2 enrich your knowledge graphs: linked data integration with pool...
Session 1.2   enrich your knowledge graphs: linked data integration with pool...Session 1.2   enrich your knowledge graphs: linked data integration with pool...
Session 1.2 enrich your knowledge graphs: linked data integration with pool...semanticsconference
 
Session 1.4 connecting information from legislation and datasets using a ca...
Session 1.4   connecting information from legislation and datasets using a ca...Session 1.4   connecting information from legislation and datasets using a ca...
Session 1.4 connecting information from legislation and datasets using a ca...semanticsconference
 
Session 1.4 a distributed network of heritage information
Session 1.4   a distributed network of heritage informationSession 1.4   a distributed network of heritage information
Session 1.4 a distributed network of heritage informationsemanticsconference
 
Session 0.0 media panel - matthias priem - gtuo - semantics 2017
Session 0.0   media panel - matthias priem - gtuo - semantics 2017Session 0.0   media panel - matthias priem - gtuo - semantics 2017
Session 0.0 media panel - matthias priem - gtuo - semantics 2017semanticsconference
 
Session 1.3 semantic asset management in the dutch rail engineering and con...
Session 1.3   semantic asset management in the dutch rail engineering and con...Session 1.3   semantic asset management in the dutch rail engineering and con...
Session 1.3 semantic asset management in the dutch rail engineering and con...semanticsconference
 
Session 1.3 energy, smart homes & smart grids: towards interoperability...
Session 1.3   energy, smart homes & smart grids: towards interoperability...Session 1.3   energy, smart homes & smart grids: towards interoperability...
Session 1.3 energy, smart homes & smart grids: towards interoperability...semanticsconference
 
Session 1.2 improving access to digital content by semantic enrichment
Session 1.2   improving access to digital content by semantic enrichmentSession 1.2   improving access to digital content by semantic enrichment
Session 1.2 improving access to digital content by semantic enrichmentsemanticsconference
 
Session 2.3 semantics for safeguarding & security – a police story
Session 2.3   semantics for safeguarding & security – a police storySession 2.3   semantics for safeguarding & security – a police story
Session 2.3 semantics for safeguarding & security – a police storysemanticsconference
 
Session 2.5 semantic similarity based clustering of license excerpts for im...
Session 2.5   semantic similarity based clustering of license excerpts for im...Session 2.5   semantic similarity based clustering of license excerpts for im...
Session 2.5 semantic similarity based clustering of license excerpts for im...semanticsconference
 
Session 4.2 unleash the triple: leveraging a corporate discovery interface....
Session 4.2   unleash the triple: leveraging a corporate discovery interface....Session 4.2   unleash the triple: leveraging a corporate discovery interface....
Session 4.2 unleash the triple: leveraging a corporate discovery interface....semanticsconference
 
Session 1.6 slovak public metadata governance and management based on linke...
Session 1.6   slovak public metadata governance and management based on linke...Session 1.6   slovak public metadata governance and management based on linke...
Session 1.6 slovak public metadata governance and management based on linke...semanticsconference
 
Session 5.6 towards a semantic outlier detection framework in wireless sens...
Session 5.6   towards a semantic outlier detection framework in wireless sens...Session 5.6   towards a semantic outlier detection framework in wireless sens...
Session 5.6 towards a semantic outlier detection framework in wireless sens...semanticsconference
 

Mais de semanticsconference (20)

Linear books to open world adventure
Linear books to open world adventureLinear books to open world adventure
Linear books to open world adventure
 
Session 1.2 high-precision, context-free entity linking exploiting unambigu...
Session 1.2   high-precision, context-free entity linking exploiting unambigu...Session 1.2   high-precision, context-free entity linking exploiting unambigu...
Session 1.2 high-precision, context-free entity linking exploiting unambigu...
 
Session 4.3 semantic annotation for enhancing collaborative ideation
Session 4.3   semantic annotation for enhancing collaborative ideationSession 4.3   semantic annotation for enhancing collaborative ideation
Session 4.3 semantic annotation for enhancing collaborative ideation
 
Session 1.1 dalicc - data licenses clearance center
Session 1.1   dalicc - data licenses clearance centerSession 1.1   dalicc - data licenses clearance center
Session 1.1 dalicc - data licenses clearance center
 
Session 1.3 context information management across smart city knowledge domains
Session 1.3   context information management across smart city knowledge domainsSession 1.3   context information management across smart city knowledge domains
Session 1.3 context information management across smart city knowledge domains
 
Session 0.0 aussenac semanticsnl-pwebsem2017-v4
Session 0.0   aussenac semanticsnl-pwebsem2017-v4Session 0.0   aussenac semanticsnl-pwebsem2017-v4
Session 0.0 aussenac semanticsnl-pwebsem2017-v4
 
Session 0.0 keynote sandeep sacheti - final hi res
Session 0.0   keynote sandeep sacheti - final hi resSession 0.0   keynote sandeep sacheti - final hi res
Session 0.0 keynote sandeep sacheti - final hi res
 
Session 1.1 linked data applied: a field report from the netherlands
Session 1.1   linked data applied: a field report from the netherlandsSession 1.1   linked data applied: a field report from the netherlands
Session 1.1 linked data applied: a field report from the netherlands
 
Session 1.2 enrich your knowledge graphs: linked data integration with pool...
Session 1.2   enrich your knowledge graphs: linked data integration with pool...Session 1.2   enrich your knowledge graphs: linked data integration with pool...
Session 1.2 enrich your knowledge graphs: linked data integration with pool...
 
Session 1.4 connecting information from legislation and datasets using a ca...
Session 1.4   connecting information from legislation and datasets using a ca...Session 1.4   connecting information from legislation and datasets using a ca...
Session 1.4 connecting information from legislation and datasets using a ca...
 
Session 1.4 a distributed network of heritage information
Session 1.4   a distributed network of heritage informationSession 1.4   a distributed network of heritage information
Session 1.4 a distributed network of heritage information
 
Session 0.0 media panel - matthias priem - gtuo - semantics 2017
Session 0.0   media panel - matthias priem - gtuo - semantics 2017Session 0.0   media panel - matthias priem - gtuo - semantics 2017
Session 0.0 media panel - matthias priem - gtuo - semantics 2017
 
Session 1.3 semantic asset management in the dutch rail engineering and con...
Session 1.3   semantic asset management in the dutch rail engineering and con...Session 1.3   semantic asset management in the dutch rail engineering and con...
Session 1.3 semantic asset management in the dutch rail engineering and con...
 
Session 1.3 energy, smart homes & smart grids: towards interoperability...
Session 1.3   energy, smart homes & smart grids: towards interoperability...Session 1.3   energy, smart homes & smart grids: towards interoperability...
Session 1.3 energy, smart homes & smart grids: towards interoperability...
 
Session 1.2 improving access to digital content by semantic enrichment
Session 1.2   improving access to digital content by semantic enrichmentSession 1.2   improving access to digital content by semantic enrichment
Session 1.2 improving access to digital content by semantic enrichment
 
Session 2.3 semantics for safeguarding & security – a police story
Session 2.3   semantics for safeguarding & security – a police storySession 2.3   semantics for safeguarding & security – a police story
Session 2.3 semantics for safeguarding & security – a police story
 
Session 2.5 semantic similarity based clustering of license excerpts for im...
Session 2.5   semantic similarity based clustering of license excerpts for im...Session 2.5   semantic similarity based clustering of license excerpts for im...
Session 2.5 semantic similarity based clustering of license excerpts for im...
 
Session 4.2 unleash the triple: leveraging a corporate discovery interface....
Session 4.2   unleash the triple: leveraging a corporate discovery interface....Session 4.2   unleash the triple: leveraging a corporate discovery interface....
Session 4.2 unleash the triple: leveraging a corporate discovery interface....
 
Session 1.6 slovak public metadata governance and management based on linke...
Session 1.6   slovak public metadata governance and management based on linke...Session 1.6   slovak public metadata governance and management based on linke...
Session 1.6 slovak public metadata governance and management based on linke...
 
Session 5.6 towards a semantic outlier detection framework in wireless sens...
Session 5.6   towards a semantic outlier detection framework in wireless sens...Session 5.6   towards a semantic outlier detection framework in wireless sens...
Session 5.6 towards a semantic outlier detection framework in wireless sens...
 

Último

Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
Accelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with PlatformlessAccelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with PlatformlessWSO2
 
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)Mark Simos
 
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sectoritnewsafrica
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfIngrid Airi González
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesMuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesManik S Magar
 
JET Technology Labs White Paper for Virtualized Security and Encryption Techn...
JET Technology Labs White Paper for Virtualized Security and Encryption Techn...JET Technology Labs White Paper for Virtualized Security and Encryption Techn...
JET Technology Labs White Paper for Virtualized Security and Encryption Techn...amber724300
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Nikki Chapple
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch TuesdayIvanti
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesThousandEyes
 
Landscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdfLandscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdfAarwolf Industries LLC
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...itnewsafrica
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
Digital Tools & AI in Career Development
Digital Tools & AI in Career DevelopmentDigital Tools & AI in Career Development
Digital Tools & AI in Career DevelopmentMahmoud Rabie
 
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...Jeffrey Haguewood
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesKari Kakkonen
 

Último (20)

Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
Accelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with PlatformlessAccelerating Enterprise Software Engineering with Platformless
Accelerating Enterprise Software Engineering with Platformless
 
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
Tampa BSides - The No BS SOC (slides from April 6, 2024 talk)
 
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
4. Cobus Valentine- Cybersecurity Threats and Solutions for the Public Sector
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdf
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesMuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
 
JET Technology Labs White Paper for Virtualized Security and Encryption Techn...
JET Technology Labs White Paper for Virtualized Security and Encryption Techn...JET Technology Labs White Paper for Virtualized Security and Encryption Techn...
JET Technology Labs White Paper for Virtualized Security and Encryption Techn...
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch Tuesday
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
 
Landscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdfLandscape Catalogue 2024 Australia-1.pdf
Landscape Catalogue 2024 Australia-1.pdf
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
Digital Tools & AI in Career Development
Digital Tools & AI in Career DevelopmentDigital Tools & AI in Career Development
Digital Tools & AI in Career Development
 
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
Email Marketing Automation for Bonterra Impact Management (fka Social Solutio...
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examples
 

Jörg Waitelonis, Henrik Jürges and Harald Sack | Don't compare Apples to Oranges - Extending GERBIL for a fine grained NEL evaluation

  • 1. Don’t compare Apples to Oranges - Extending GERBIL for a fine grained NEL Evaluation Jörg Waitelonis, Henrik Jürges, Harald Sack Hasso-Plattner-Institute for IT-Systems Engineering, University of Potsdam Semantics 2016, Leipzig, Germany, September 12-15th, 2016
  • 2. Agenda 1. NEL and NEL evaluation 2. Dataset properties and evaluation drawbacks 3. Extending GERBIL ● Building conditional datasets ● Measure dataset characteristics 1. Results 2. Demonstration 3. Summary & Future work
  • 3. Named Entity Linking (NEL) Chart 3 … Armstrong …
  • 4. Named Entity Linking (NEL), Principle Chart 4 “Armstrong landed on the moon.” Candidates: dbr:Neil_Armstrong dbr:Lance_Armstrong dbr:Louis_Armstrong …. Candidates: dbr:Moon dbr:Lunar …. Correct entities Entity mention with surface form String Distance Link Analysis Vector Space Fuzzy String Matching Conditional Random Fields Random Forest RankSVM Learning to Rank Surface Aggregation Word Embeddings Context Similarity Matching 1. Tokenize text 2. Find candidates in KB 3. Score candidates with a magic algorithm and select the best one KEA Wikifier ● Algorithm only approximates correct entities ● Need for verification and testing
  • 5. ● Dataset consists out of: ■ Documents (String/Sentences) ■ Annotations (ground truth) Named Entity Linking, Evaluation Chart 5 ACE2004 AIDA/CoNLL DBpedia Spotlight IITB KORE50 MSNBC Micropost2014 N3-RSS-500 WES2015 N3-Reuters-128 ● Traditional measures are: ■ Precision: defines how well an annotator works ■ Recall: defines how complete the results are ■ F1-measure: harmonic medium between precision and recall ■ And more, cf. Rizzo et al. [1]
  • 6. ● GERBIL - a general entity annotation system (AKSW Leipzig), cf. Usbeck et. al [2] ● Used for testing/optimizing/benchmarking annotators ● Neat Webinterface ● 13 Annotators / 20 datasets ● F-measure to rough for detailed evaluation ● Developer need dataset insights Named Entity Linking, Benchmarking Chart 6 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 7. ● Size of a datasets ● Amount of annotations/documents/words ● What types of entities are used? E.g. persons, places, events, …. ● Are there documents without annotations? E.g. Microposts 2014 ● What sort of popularity have the entities? E.g. PageRank, Indegree ● How ambiguous are the entities and surface forms? ● How divers are the entities and surface forms? ● …. Cf. van Erp et al. [3] Properties of Datasets Chart 7 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 8. ● How does the dataset characteristics influence the evaluation results? ● How does the popularity of entities influence the evaluation results? ● How can a general dataset be used for domain specific NEL tools? ● How can datasets be compared? Is there something like a general difficulty? ● Limited comparability between benchmark results ● Penalization of good annotators with inappropriate datasets Cf. van Erp et al. [3] Research Questions and Drawbacks Chart 8 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 9. ● Approach for a solution: ● Adjustable filter system for GERBIL ● Expose dataset characteristics ● Datasets and annotators added at runtime are also included ● Visualize the results Extending GERBIL Chart 9 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 10. Extending GERBIL, Conditional Datasets Chart 10 Dataset Type and popularity specific datasets Annotator Documents Annotator results Benchmark results Evaluate each specific dataset and result PR(e) > t PR(e) > t PR(e) > t rdf:type rdf:type rdf:type rdf:type rdf:type rdf:type
  • 11. Results, Types Chart 11 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 13. Extending GERBIL, Not Annotated Documents ● Not annotated documents: shows the relative amount of empty documents within a datasets ● Only affects if annotators searches entity mentions by themselves Chart 13 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 14. Extending GERBIL, Density Chart 14 ● Density: shows the relation between number of annotations and words in the document ● Only affects if annotators searches entity mentions by themselves Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 15. Extending GERBIL, Likelihood of Confusion ● Likelihood of Confusion (Level of Ambiguity) ● True measures are unknown due to missing exhaustive collections ● Rough overview how difficult to disambiguate Chart 15 Entities Surface Forms Tegel TXL Bruce Otto Lilienthal Bruce Lee Bruce Willis Airport Tegel Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation synonyms
  • 16. Results, Likelihood of Confusion Chart 16 Entities ● A high red bar indicates an entity has a high amount of homonyms ● A high blue bar indicates a surface form has a high amount of synonyms Surface Forms
  • 17. Extending GERBIL, Dominance of Entities Chart 17 Bruce Bruci Bruce Willis Testdata Vocabulary dominance(e)= e(t)/e(v) ● Expresses the relation between used words and all words ● True measures unknown ● High rates prevents overfitting ● Prevents repetition of surface forms dbr:Bruce_Willis Bruce Walter Willis
  • 18. Extending GERBIL, Dominance of Surface Forms Chart 18Chart 18 dbr:Irene_Angelina dbr:Angelina_Jordan dbr:Angelina_Jolie Vocabulary dominance(s)= s(t)/s(v) ● Expresses the relation between used mentions and all mentions ● True measures unknown ● High rates prevents overfitting ● Indicates how context dependent a disambiguation is Testdata Angelina
  • 19. Results, Dominance Chart 19 ● Blue bar indicates that for a entity a variety of surface forms is used ● Red bar indicates how context dependent the disambiguation of an surface form is Dominance of surface forms Dominance of entities
  • 20. Demo ● http://gerbil.s16a.org/ ● https://github.com/santifa/gerbil/ Chart 20 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 21. ■ Summary: □ Implemented a domain specific filter system □ Measure dataset characteristics □ Annotator results are nearly the same on entities of different popularity □ Enable specific analyses and optimization of annotators □ Enable users to select the tools that performs best for a specific domain ■ Future work: □ Keep up with GERBIL development, increase performance □ More measurements, e. g. max_recall □ Dataset remixing ≙ assemble new customized datasets – E. g. Unpopular companies Summary & Future work Chart 21 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation
  • 22. [1] Giuseppe Rizzo, Amparo Elizabeth Cano Basave, Bianca Pereira, and Andrea Varga. Making Sense of Microposts (#Microposts2015) Named Entity rEcognition and Linking (NEEL) Challenge. In 5th Workshop on Making Sense of Microposts (#Microposts2015), pages 44–53. CEUR-WS.org, 2015 [2] M. Röder, R. Usbeck, and A.-C. Ngonga Ngomo. Gerbil’s new stunts: Semantic annotation benchmarking improved. Technical report, Leipzig University, 2016 [3] M. van Erp, P. Mendes, H. Paulheim, F. Ilievski, J. Plu, G. Rizzo, and J. Waitelonis. Evaluating entity linking: An analysis of current benchmark datasets and a roadmap for doing a better job. In Proc. of 10th edition of the Language Resources and Evaluation Conference, Portoroz, Slovenia, 2016. References Chart 22 Henrik Jürges Semantics 2016 Leipzig, Germany Don’t compare Apples to Oranges: Extending GERBIL for a fine grained NEL Evaluation

Notas do Editor

  1. Hello and welcome to this talk. Thank you all very much for coming today. My name is Henrik Jürges and I’m student co worker at the semantic web technologies research group at the hpi This is joint work with Jörg Waitelonis and Harald Sack The title of my presentation is don’t compare apple to oranges - extending gerbil for a fine grained nel evaluation And as the name suggest my purpose is to give you a brief overview of our approach to follow new trends in named entity linking (So I’m hoping to cover three points. Firstly I will introduce GERBIL and it’s disadvantages to cover up new trends, after that we will look At the work we have done to lay a foundation for future work in this direction and finally I will present some results.)
  2. The presentation is structured as follows: First i give you a brief overview about named entity linking and its evaluation After that i state some questions regarding current data sets and evaluation methods Then I show you our approach to deal with these questions and present you some results we got Finally, i sum up this presentation and give you quick view into future work
  3. A more complete example is the next sentence: Armstrong landed on the moon Generally the same three steps are applied by almost every nel tool or so called annotator The text is split into tokens by white spaces and important entity mentions are located. These are called surface forms After that possible candidates are searched from a formal knowledge-base like dbpedia e.g. by comparing the surface forms and entity labels Then some magic algorithm ranks these candidates. As you can see there are many possible algorithms from easy ones like string distance to Complex ones like random forest. But these are all implemented by various tools you can see on the right side. The highest ranked candidate will hopefully our correct entity This will lead us to the evaluation of named entity linking Let me show you a more enlightening example of named entity linking First we have some text fragment, mostly a sentence or some larger text. Here the phrase “Armstrong landed on the moon.” is your fragment. Then we have the application fulfilling the task of named entity linking Mostly all named entity linking tools or so called annotators have three basic stages: Firstly they tokenize the text into words. After that they search for candidates in some formal knowledge-base like dbpedia. Finally they apply a magic algorithm on these candidates, finding the best candidate. Here we should distinguish between two common task in named entity linking. One task only disambiguates provided entity mentions with the help of the text fragment, Called disambiguate to knowledge-base The other task provides only the text fragment leaving the annotator to find expected entity mentions and disambiguate them. This task is called annotate to knowledge-base and Encapsulates the former task. Ok, taking this slow, we have our text fragment and tokenize the text searching for entity mentions. We found two possible entity mentions: armstrong and the moon. The textual fragments for the two mentions are called surface forms. Which are indeed just words who are syntactically right. After that we search one or multiple formal knowledge-base for the possible candidates for each surface form. Leading us to the candidates Neil Armstrong, Lance Armstrong, Louis Armstrong and so on. Having all candidates together the annotator uses some magic algorithm for scoring the candidates. These algorithms are varying in their complexity from easy ones like string distance to more complex ones like random forest or context similarity matching. Glad for us there are many annotators implementing these algorithms. Some of them are shown at the right side. By scoring the candidates, the candidate with the highest score might be our correct entity we’re searching for. This will lead us to the entity Neil armstrong and the moon for our entity mentions.
  4. Since these algorithm only approximate the correct entity, we need evaluation to verify and test the results To do so we need data sets which are the models of our expectation A data sets contains two things, first the documents or text fragments and as second the expected annotations for these documents which is called ground truth For the evaluation we need some measures which are withdrawn from other research areas The most common measures are precision which defines how well the results are and recall which defines how complete the results are. And to combine these the f one measure gives an overview about the quality of the annotators results Since we have the matrix describing our expectations, we need something that represents our expectation. For this we use a data sets. A data sets models our expectation within two aspects. Firstly it contains the documents or text fragments which we use for evaluation. These documents are mostly string which are one or more sentences. Secondly a set of annotations so called goldstandard or ground truth which is our expectation for the annotator results for some documents The ground truth is either hand-crafted by multiple researchers or taken from annotators and correct later Some of the known data sets are mentioned at the right With the help of our expectation matrix we can now measure the amount of correct, missing and false results. But the values are a bit hard to compare and are without deeper meaning, so some measure a borrowed from other research areas. I will present the three most common here which are all based on our expectation matrix. First the precision defines how well an annotator works. Secondly the recall describes how complete the results are. And lastly the f one measure is the harmonic medium between precision and recall. It gives an overview about the quality an annotator produces.
  5. But doing this by hand is a quite error-prone and annoying task, so benchmarking and evaluation system evolved over time and the successor of all is GERBIL GERBIL is work done by our colleagues from AKSW Leipzig and is a general entity annotation system which can be used for testing optimizing and benchmarking Annotators. Provides a nice web interface for configuring, benchmarking and visualizing data sets and annotators At the time there are 13 annotators and 20 data sets provided but new data sets and annotators can be added at runtime The results are presented as a spider diagram. As you see in the spider diagram every dot is the f1 measure for an annotator and a specific data sets that owns this line. But when see some further there are minor or major changes for an annotator between the data sets which leads us to the question why? Some changes a relatively clear like the micropost data sets contains short sentences with only a little bit of context but other changes are not quite clear. This is a major drawback when developing an annotator and you don’t know all characteristics for each data set With all the measures and data sets we have a foundation for the evaluation, but for now this is all done by hand and the results Are not comparable across single evaluations or between annotators. So GERBIL has evolved which is a general entity annotation system. It can be used for testing, optimizing and comparing annotator.. It provides a neat web interface and a possibility to add new data sets and annotators at runtime. Entwicklerperspektive, ich brauche die eigenschaften der datensätze um annotator zu verbessern
  6. This lead us to the question of measuring data sets properties There are basic properties like the size of a data sets, the amount of annotation or documents in a data set And some advanced properties like what types of entities are used, are there documents without annotations, Which popularity the entities have. And also how ambig or divers are the used entities and surface forms?
  7. Considering these properties and the whole evaluation process we came up with some questions and drawbacks in regard of named entity linking evaluation Can we show the influence of some data set characteristics Are annotators better on popular entities? Can we use the existing data sets for annotators that are focused on some domain especially entity types. Is there a way to compare data sets or something like a difficulty level. In general we found that results are less comparable between data sets and some quite good annotator could be penalized if using in conjunction with inappropriate data sets. So to sum these problems up: We have a general benchmarking approach against domain specific annotators with focus on geotagging, person, organization’s, tweets and so on The data sets are annotated sparsely with old knowledge-bases in mind but modern nel tools are more precise and the knowledge-bases are more complete The assumption was that all data sets have the same difficulty level but they don’t All this led to results which are not comparable and which penalize good annotators Besides that there are more problems growing out of this few: How we get and manage all the domain specific data sets? How we could define a difficulty level? And how we can keep up with upcoming trends?
  8. To tackle these questions we have done multiple things We introduced an arbitrary filter system which can be used for domain specific evaluation We implemented some characteristics for data sets Booth things are automatically applied and there are expendable And we provide some visual graphics for the results
  9. To provide a domain specific evaluation we build an arbitrary filter system. A general schematic can be seen on the right side. As you can see, we get a list of annotations. First we unwrap them leaving only the IRIs of the entities, then we clean the List from IRIs which are not complained with the standard For performance issues we cache the results and split them into chunks At this time we implemented two basic filters one dedicated for SPARQL queries and one for Popularity. A configuration example is shown on the left side. Every filter takes a name and a backend service like dbpedia or a file or something else The real filter is here a sparql query which returns all entity links which are persons. The doublecross is replaced with the real links And to cover some issues handling the amount of data, the filter can define a chunk size. The popularity filter follows the same conventions only using another service and filter query.
  10. We came up with two measures for sparse data sets. The first one show the relative amount of empty documents within a data sets. Meaning that one or more documents don’t have any annotations at all. And the second one is called Missing annotations which is a little bit misleading so density is the better name. Since we can not measure what is missing we decided to measure how many expectations are in a document with respect to all possible expectations for a document also known As words. It is quite important to mention that these measures only affect a certain subtask in named entity linking. If an annotator only gets a text fragment and searches for entities by themselve, it is possible to find entity mentions in not annotated documents or new one in data sets with a low density. This leads to an penalization since the annotator results are taken as false positives although they could be right.
  11. The ambiguity can be described in booth directions for either surface forms or entities For example the entity airport tegel can be described through three surface which makes them all synonyms The other way round the surface form bruce is linked to two entities which makes them homonyms Since no exhaustive collection of relations between surface forms and entities exists the true measures remain unknown But the level of ambiguity gives a rough extent about how hard to disambiguate is a dataset.