CHEAP Call Girls in Rabindra Nagar (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
Reproducibility for IR evaluation
1. Department of Information Engineering
University of Padua, Italy
Gianmaria Silvello
@giansilv
Reproducibility for IR Evaluation
2. slideReproducibility for IR EvaluationG. Silvello
IR Evaluation Initiatives
2
Evaluation in IR is often conducted in large, shared,
international campaigns
FIRE
3. slideReproducibility for IR EvaluationG. Silvello
IR Evaluation Initiatives
3
Organizer
Assessor
Par.cipant
Organizer
Assessor
Visitor,
Par.cipant
and
Organizer
Visitor,
Par.cipant
and
Organizer
Visitor,
Par.cipant
and
Organizer
Prepara.on of
Documents
Crea.on
of Topics
Experiment
Submission
Crea.on of
Pools
Relevance
Assessment
Performance
Measures
Scien.fic
Produc.on
Data
Informa.on
Knowledge Wisdom
Sta.s.cal
Analyses
4. slideReproducibility for IR EvaluationG. Silvello
IR Evaluation Initiatives
3
Organizer
Assessor
Par.cipant
Organizer
Assessor
Visitor,
Par.cipant
and
Organizer
Visitor,
Par.cipant
and
Organizer
Visitor,
Par.cipant
and
Organizer
Prepara.on of
Documents
Crea.on
of Topics
Experiment
Submission
Crea.on of
Pools
Relevance
Assessment
Performance
Measures
Scien.fic
Produc.on
Data
Informa.on
Knowledge Wisdom
Sta.s.cal
Analyses
We have shared
experimental
collections and we
perform statistical
validation.
But, are we done?
5. slideReproducibility for IR EvaluationG. Silvello
IR Evaluation Initiatives
3
Organizer
Assessor
Par.cipant
Organizer
Assessor
Visitor,
Par.cipant
and
Organizer
Visitor,
Par.cipant
and
Organizer
Visitor,
Par.cipant
and
Organizer
Prepara.on of
Documents
Crea.on
of Topics
Experiment
Submission
Crea.on of
Pools
Relevance
Assessment
Performance
Measures
Scien.fic
Produc.on
Data
Informa.on
Knowledge Wisdom
Sta.s.cal
Analyses
Multiple targets for reproducibility:
experimental collections
system runs
meta-evaluation studies
7. slideReproducibility for IR EvaluationG. Silvello
The Format Babele
4
This situation hampers:
- automatic management
- interpretability
- reproducibility
- ease of (re-)use
- take-up from new comers
303#0#APW19980609.1531#21
303#0#APW19980610.1778#12
303#0#APW19980715.1061#23
303#0#APW19980910.1078#04
5
1#0#clueweb095en0120513520479#06
1#1#clueweb095en0120513520479#07
1#2#clueweb095en0120513520479#08
9
101#0#clueweb095en0047533520039#110
101#0#clueweb095en0004566509322#211
101#0#clueweb095en0033530508382#012
101#0#clueweb095en0000545505740#5213
101#0#clueweb095en0020592511795#114
15
20002#0#clueweb095en0006585533170#1#1#10.516
20004#0#clueweb095en0005528520976#1#1#10.517
20006#0#clueweb095en0010507521538#1#1#10.518
19
ad-hoc
diversity
ad-hoc with grades
relevance feedback
8. slideReproducibility for IR EvaluationG. Silvello
The Format Babele
4
This situation hampers:
- automatic management
- interpretability
- reproducibility
- ease of (re-)use
- take-up from new comers
303#0#APW19980609.1531#21
303#0#APW19980610.1778#12
303#0#APW19980715.1061#23
303#0#APW19980910.1078#04
5
1#0#clueweb095en0120513520479#06
1#1#clueweb095en0120513520479#07
1#2#clueweb095en0120513520479#08
9
101#0#clueweb095en0047533520039#110
101#0#clueweb095en0004566509322#211
101#0#clueweb095en0033530508382#012
101#0#clueweb095en0000545505740#5213
101#0#clueweb095en0020592511795#114
15
20002#0#clueweb095en0006585533170#1#1#10.516
20004#0#clueweb095en0005528520976#1#1#10.517
20006#0#clueweb095en0010507521538#1#1#10.518
19
ad-hoc
diversity
ad-hoc with grades
relevance feedback
We need:
to agree on a common data model which
allows for extension
to provide the basic experimental data
with proper metadata (descriptive,
administrative, copyright, ...)
9. slideReproducibility for IR EvaluationG. Silvello
Referenceability and Traceability
5
- Explanation of experimental data is usually reported in
scientific papers that do not provide direct links to them
- the may be referred to in many different ways within the same
paper (experiment id, system version, participant id, …)
- It is often difficult to exactly know which data have been
used in a paper and have access to them
- It is ever more difficult to exactly know the performed data
cleaning and processing operations
[Ferro, 2016]
10. slideReproducibility for IR EvaluationG. Silvello
Referenceability and Traceability
5
- Explanation of experimental data is usually reported in
scientific papers that do not provide direct links to them
- the may be referred to in many different ways within the same
paper (experiment id, system version, participant id, …)
- It is often difficult to exactly know which data have been
used in a paper and have access to them
- It is ever more difficult to exactly know the performed data
cleaning and processing operations
We need:
to have the possibility of citing experimental
data in our papers as any other references
and to link the data with the claims in the
papers
to make our papers actionable and
executable providing access to the
mentioned experimental data
[Ferro, 2016]
11. slideReproducibility for IR EvaluationG. Silvello
The DIRECT Experience
6
BIBLIOGRAPHICAL
EXPERIMENT
VISUAL
ANALYTICS
EVALUATION
ACTIVITY
EXPERIMENTAL
COLLECTION
RESOURCE
MANAGEMENT
MEASUREMENT
METADATA
http://direct.dei.unipd.it/
http://lod-direct.dei.unipd.it/
[Agosti et al., 2012]
12. slideReproducibility for IR EvaluationG. Silvello
LOD DIRECT
7
Jussi
Karlgren
Link ims:relation
ims:has-source
ims:has-target
is-expert-in
Reputation
Management
0.46 0.84
ims:score ims:backward-score
CLEF2012wn-
RepLab-
KarlgrenEtAl
2012
Link
ims:has-source
ims:has-target
ims:relation
feature
0.53 0.87
ims:score
ims:backward-score
Profiling Reputation of Corporate
Entities in Semantic Space
ims:title
dbpedia.or
g/resource/
Reputation_
manageme
nt
owl:sameAs
dbpedia.or
g/resource/
Information
_
retrieval
Link
ims:has-source
ims:has-target
ims:relation
0.42 0.23
ims:score ims:backward-score
Information
Retrieval
owl:sameAs
swrc:has-author
dblp.l3s.de/d2r/
resource/
publications/
conf/clef/
KarlgrenSOEH1
2
owl:sameAs
dblp.l3s.de/
d2r/resource/
authors/
Jussi_Karlgre
n owl:sameAs
RepLab
2012
CLEF
2012
profiling
_kthgavagai
_1
Measure
0.77 ims:score
Effectiveness
Accuracy
ims:refersTo
ims:submittedTo
ims:isPartOf
ims:evaluates
ims:isEvaluatedBy
ims:assignedTo
ims:measuredBy
[Silvello et al., 2016]
16. slideReproducibility for IR EvaluationG. Silvello
Actionable Papers
9
<a href=”http://direct.dei.unipd.it/user/UPV”>UPV</a>
17. slideReproducibility for IR EvaluationG. Silvello
Actionable Papers
9
<a href=”http://direct.dei.unipd.it/experiment/
EXP_UKB_WN100”>EXP_UKB_WN100</a>
18. slideReproducibility for IR EvaluationG. Silvello
Actionable Papers
9
<a href=”http://direct.dei.unipd.it/estimate/
017c333a-4b7c-4267-926d-f15fe3554efd”>51.61%</a>
19. slideReproducibility for IR EvaluationG. Silvello
Actionable Papers
9
<img src=”http://direct.dei.unipd.it/visualization/
017c333a-4b7c-4267-926d-f15fe3554efd/snapshot/
177bcef2-00a0-4f59-b781-f285610f1c6f”/>
20. slideReproducibility for IR EvaluationG. Silvello
Reproducibility is tied to data citation
10
Being able to uniquely identify data (e.g., DOI, URI) is fundamental, but it is not enough
- We need to:
- automatically generate pertinent, consistent and complete human-
and machine-readable citation snippets
- define tool to make data citation easy: click, generate, copy and
paste
- develop citation systems which require low or no effort to data
creators/curators and low or no modification to the actual data being
cited
- make persistent data citations
[Silvello&Ferro, 2016]
21. slideReproducibility for IR EvaluationG. Silvello
Data Citation is a Computational Problem
11
- Identity
- Completeness
- Fixity
- Validity
[Buneman, Davidson,
Frew, 2016]
The four main computational
issues of data citation
22. slideReproducibility for IR EvaluationG. Silvello
Towards a General Data Citation System
12
The identity+completeness issues
To identify and generate a citation for a single resource
<Iuphar>
<name>IUPHAR-DB </name>
<citation>Rule0</citation>
[...]
<gpcr>
<name>G protein-coupled receptors</name>
<citation>Rule1</citation>
[...]
<family>
<id>29</id>
<name>Glucagon receptor family</name>
<citation>Rule2</citation>
<receptor>
<id>247</id>
<name>GHRH</name>
[...]
<agonists>
<ligand>
[...]
</ligand>
</agonists>
[...]
</receptor>
[...]
</family>
[...]
</gpcr>
<ionchannels>
[...]
</ionchannels>
</iuphar>
iuphar[name=$.d,url=$.u, version=$.v]
iuphar[]/gpcr[name=$.n]
iuphar[]/gpcr[]/family[name=$.f,id=$.i]
/contributors[]/contributor[name=$?c]
{database=$d, version=$v, contributors=$c, db-family=$n, family=$f, idFamily=$i}
Rules:
The citation that gets generated (example):
{ database=IUPHAR-DB: the IUPHAR database || url=http://www.iuphar-db.org/ || version=15 ||
dbFamily=G protein-coupled receptors || family=Glucagon receptor family || idFamily=29 || contributor=
{Laurence J. Miller;;Daniel J. Drucker;;[...];;Rebecca Hills;;}}
The rules are recursively
processed by the system and
then transformed into a
conjunction of XPaths.
The interpretation of the XPaths
generates the citation.
Instantiation of the variables:
The first rule interpreted by the
system
The second rule interpreted by
the system
The third rule interpreted by
the system
[Buneman&Silvello, 2010]
Rule-based system
for hierarchical data
23. slideReproducibility for IR EvaluationG. Silvello 13
Towards a General Data Citation System
The identity+completeness issues
To identify and generate a citation for a single resource
[Silvello, 2016]
Learning to cite framework
for hierarchical data
Human-Readable
Citations
XML Files
Collection
Training Data
Learner
Citation
Model
Citation
System
Citation
XPath
XML File
Test Data
Machine-Readable
Citation
Human-Readable
Citation
Output Reference
1
2
3
4
5 6
24. slideReproducibility for IR EvaluationG. Silvello 14
Towards a General Data Citation System
The identity+completeness issues (+ fixity)
To identify and generate a citation for a single resource
[Alawini, Chen,
Davidson & Silvello,
in preparation]
View+rule based system
for RDF datasets
e1
e2 e3
e4 e5
e6
e7
e8
e9
e10
pypz
pz
py
pz
py
px
px
px
px
py
py
pz
VSW(e1)
Resource to be cited: e1 check type
citation query
parametrized by e1 CSW(e1,s,v,d,t,o,u)
Citation
Function
{eagle-id: “eagle-id: e1'',
name: `”Significance Tester'',
developers: {“Grant, G.'', “Lazar, M.l'', “Manduchi, E.''},
url: “http://www.cbil.upenn.edu/STAR/ '' }
Final citation
RDF
Citation
Model
eagle-i id
Citation
Formatter
machine-readable
citation (JSON)
human-readable
citation
eagle-i
triple store
eagle-iV
versioning
system
26. slideReproducibility for IR EvaluationG. Silvello 16
Towards a General Data Citation System
The identity+completeness issues
To identify and generate a citation for a multiple resources
[Davidson, Deutch, Milo,Silvello, 2017]
View-based model
for relational databases
Query
Rewriting
Function
Database Views
V
Specification
Language
Query
Q
Database
D
Citation
Policies
q1
q2
qn
.
.
.
Preference
Model
Citation
Function
Set of best
rewritings
Citation Queries
CQ
c1
c2
cm
.
.
.
Aggregation
Function
Citation
C
1
2 3 4 5
Citation Views
Citation Checking Mechanism
6
27. slideReproducibility for IR EvaluationG. Silvello
Conclusions
17
- Reproducibility is a fundamental topic for science
- Information retrieval evaluation is a challenging domain
- Data Citation is a complex and open problem
- new models of citations
- computational solutions
- intrinsically related to reproducibility
28. slideReproducibility for IR EvaluationG. Silvello
References
18
[Agosti et al., 2012] Agosti, M., Di Buccio, E., Ferro, N., Masiero, I., Peruzzo, S., and Silvello, G. (2012).
DIRECTions: Design and Specification of an IR Evaluation Infrastructure. In Proceedings of the Third International
Conference of the CLEF Initiative (CLEF 2012). LNCS 7488, Springer, Heidelberg, Germany.
[Buneman et al., 2016] Buneman, P., Davidson, S. B., and Frew, J. (2016). Why data citation is a computational
problem. Communications of the ACM (CACM), 59(9):50–57.
[Buneman and Silvello, 2010] Buneman, P. and Silvello, G. (2010). A Rule-Based Citation System for Structured
and Evolving Datasets. IEEE Data Eng. Bull., 33(3):33–41.
[Davidson et al., 2017] Davidson, S. B., Deutch, D., Tova, M. and Silvello, G. (2017). A Model for Fine-Grained
Data Citation. In 8th Biennial Conference on Innovative Data Systems Research (CIDR 2017).
[Ferro, 2016] Ferro, N. (2016). Reproducibility Challenges in Information Retrieval Evaluation. ACM Journal of Data
and Information Quality (JDIQ), to appear.
[Silvello, 2015] Silvello, G. (2015). A Methodology for Citing Linked Open Data Subsets. D-Lib Magazine, 21(1/2).
[Silvello, 2016] Silvello, G. (2016). Learning to Cite Framework: How to Automatically Construct Citations for
Hierarchical Data. Journal of the American Society for Information Science and Technology (JASIST), in print:1–28.
[Silvello et al., 2016] Silvello, G., Bordea, G., Ferro, N., Buitelaar, P., and Bogers, T. (2016). Semantic
Representation and Enrichment of Information Retrieval Experimental Data. International Journal on Digital Libraries
(IJDL), in press:1–28.
[Silvello and Ferro, 2016] Silvello, G. and Ferro, N. (2016). ”Data Citation is Coming”. Introduction to the special
issue on data citation. Bulletin of IEEE Technical Committee on Digital Libraries, Special Issue on Data Citation,
12(1):1–5.