Elenco (non esaustivo) di pubblicazioni prodotte dai membri del laboratorio
2020
Amigó, Enrique; Gonzalo, Julio; Mizzaro, Stefano; Carrillo-de-Albornoz, Jorge
An Effectiveness Metric for Ordinal Classification: Formal Properties and Experimental Results Proceedings Article
In: Jurafsky, Dan; Chai, Joyce; Schluter, Natalie; Tetreault, Joel (Ed.): Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 3938–3949, Association for Computational Linguistics, Online, 2020.
@inproceedings{amigo-etal-2020-effectiveness,
title = {An Effectiveness Metric for Ordinal Classification: Formal Properties and Experimental Results},
author = {Enrique Amigó and Julio Gonzalo and Stefano Mizzaro and Jorge Carrillo-de-Albornoz},
editor = {Dan Jurafsky and Joyce Chai and Natalie Schluter and Joel Tetreault},
url = {https://aclanthology.org/2020.acl-main.363},
doi = {10.18653/v1/2020.acl-main.363},
year = {2020},
date = {2020-01-01},
booktitle = {Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics},
pages = {3938–3949},
publisher = {Association for Computational Linguistics},
address = {Online},
abstract = {In Ordinal Classification tasks, items have to be assigned to classes that have a relative ordering, such as ``positive'', ``neutral'', ``negative'' in sentiment analysis. Remarkably, the most popular evaluation metrics for ordinal classification tasks either ignore relevant information (for instance, precision/recall on each of the classes ignores their relative ordering) or assume additional information (for instance, Mean Average Error assumes absolute distances between classes). In this paper we propose a new metric for Ordinal Classification, Closeness Evaluation Measure, that is rooted on Measurement Theory and Information Theory. Our theoretical analysis and experimental results over both synthetic data and data from NLP shared tasks indicate that the proposed metric captures quality aspects from different traditional tasks simultaneously. In addition, it generalizes some popular classification (nominal scale) and error minimization (interval scale) metrics, depending on the measurement scale in which it is instantiated.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Mizzaro, Stefano
On the nature of information access evaluation metrics: a unifying framework Journal Article
In: Information Retrieval Journal, vol. 23, no 3, pp. 318–386, 2020, ISSN: 1573-7659.
@article{Amigo2020,
title = {On the nature of information access evaluation metrics: a unifying framework},
author = {Enrique Amigó and Stefano Mizzaro},
url = {https://doi.org/10.1007/s10791-020-09374-0},
doi = {10.1007/s10791-020-09374-0},
issn = {1573-7659},
year = {2020},
date = {2020-01-01},
journal = {Information Retrieval Journal},
volume = {23},
number = {3},
pages = {318–386},
abstract = {We provide a uniform, general, and complete formal account of evaluation metrics for ranking, classification, clustering, and other information access problems. We leverage concepts from measurement theory, such as scale types and permissible transformation functions, and we capture the nature of evaluation metrics in many tasks by two formal definitions, which lead to a distinction of two metric/tasks families, and provide a comprehensive classification of the tasks that have been proposed so far. We derive some theorems to analyze the suitability (or otherwise) of some common metrics. Within our model we can derive and explain the theoretical properties and drawbacks of the state of the art metrics for multiple tasks. The main contributions of this paper are that, differently from previous studies, the formalization is well grounded on a solid discipline, it is general as it can take into account most effectiveness metrics as well as most existing tasks, and it allows to derive important consequences on metrics and their limitations.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Amigó, Enrique; Fang, Hui; Mizzaro, Stefano; Zhai, Chengxiang
Axiomatic thinking for information retrieval: introduction to special issue Journal Article
In: Information Retrieval Journal, vol. 23, no 3, pp. 187–190, 2020, ISSN: 1573-7659.
@article{Amigo2020b,
title = {Axiomatic thinking for information retrieval: introduction to special issue},
author = {Enrique Amigó and Hui Fang and Stefano Mizzaro and Chengxiang Zhai},
url = {https://doi.org/10.1007/s10791-020-09376-y},
doi = {10.1007/s10791-020-09376-y},
issn = {1573-7659},
year = {2020},
date = {2020-01-01},
journal = {Information Retrieval Journal},
volume = {23},
number = {3},
pages = {187–190},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2019
Roitero, Kevin; Soprano, Michael; Mizzaro, Stefano
Bias and Fairness in Effectiveness Evaluation by Means of Network Analysis and Mixture Models Proceedings Article
In: CEUR Workshop Proceedings, pp. 2, CEUR-WS, 2019.
@inproceedings{bias-fairness-19,
title = {Bias and Fairness in Effectiveness Evaluation by Means of Network Analysis and Mixture Models},
author = {Kevin Roitero and Michael Soprano and Stefano Mizzaro},
year = {2019},
date = {2019-10-14},
booktitle = {CEUR Workshop Proceedings},
volume = {2441},
pages = {2},
publisher = {CEUR-WS},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Han, L; Roitero, K; Gadiraju, U; Sarasua, C; Checco, A; Maddalena, E; Demartini, G
The Impact of Task Abandonment in Crowdsourcing Journal Article
In: IEEE Transactions on Knowledge & Data Engineering, no 01, pp. 1-1, 2019, ISSN: 1558-2191.
@article{8873609,
title = {The Impact of Task Abandonment in Crowdsourcing},
author = {L Han and K Roitero and U Gadiraju and C Sarasua and A Checco and E Maddalena and G Demartini},
doi = {10.1109/TKDE.2019.2948168},
issn = {1558-2191},
year = {2019},
date = {2019-10-01},
journal = {IEEE Transactions on Knowledge & Data Engineering},
number = {01},
pages = {1-1},
publisher = {IEEE Computer Society},
address = {Los Alamitos, CA, USA},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Kevin,; J, Culpepper Shane; Mark, Sanderson; Falk, Scholer; Roitero, Mizzaro Stefano
Fewer topics? A million topics? Both?! On topics subsets in test collections Journal Article
In: Information Retrieval Journal, 2019, ISSN: 1573-7659.
@article{Roitero2019ffew,
title = {Fewer topics? A million topics? Both?! On topics subsets in test collections},
author = {Kevin and Culpepper Shane J and Sanderson Mark and Scholer Falk and Mizzaro Stefano Roitero},
url = {https://doi.org/10.1007/s10791-019-09357-w},
doi = {10.1007/s10791-019-09357-w},
issn = {1573-7659},
year = {2019},
date = {2019-05-08},
journal = {Information Retrieval Journal},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Roitero, Kevin; Mizzaro, Stefano; Soprano, Michael
Bias and Fairness in Effectiveness Evaluation by Means of Network Analysis and Mixture Models Proceedings Article
In: Proceedings of the 10th Italian Information Retrieval Workshop, Padova, Italy, September 16-18, 2019., pp. 6–7, 2019.
@inproceedings{DBLP:conf/iir/RoiteroMS19,
title = {Bias and Fairness in Effectiveness Evaluation by Means of Network
Analysis and Mixture Models},
author = {Kevin Roitero and Stefano Mizzaro and Michael Soprano},
url = {http://ceur-ws.org/Vol-2441/paper4.pdf},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 10th Italian Information Retrieval Workshop, Padova,
Italy, September 16-18, 2019.},
pages = {6--7},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soprano, Michael; Roitero, Kevin; Mizzaro, Stefano
HITS Hits Readersourcing: Validating Peer Review Alternatives Using Network Analysis Proceedings Article
In: Proceedings of the 4th Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries (BIRNDL 2019) co-located with the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2019), Paris, France, July 25, 2019., pp. 70–82, 2019.
@inproceedings{DBLP:conf/sigir/SopranoRM19,
title = {HITS Hits Readersourcing: Validating Peer Review Alternatives Using
Network Analysis},
author = {Michael Soprano and Kevin Roitero and Stefano Mizzaro},
url = {http://ceur-ws.org/Vol-2414/paper7.pdf},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 4th Joint Workshop on Bibliometric-enhanced Information
Retrieval and Natural Language Processing for Digital Libraries (BIRNDL 2019) co-located with the 42nd International ACM SIGIR Conference
on Research and Development in Information Retrieval (SIGIR 2019),
Paris, France, July 25, 2019.},
pages = {70--82},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Zampieri, Fabio; Roitero, Kevin; Culpepper, Shane J; Kurland, Oren; Mizzaro, Stefano
On Topic Difficulty in IR Evaluation: The Effect of Systems, Corpora, and System Components Proceedings Article
In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2019, Paris, France, July 21-25, 2019., pp. 909–912, 2019.
@inproceedings{DBLP:conf/sigir/ZampieriRCKM19,
title = {On Topic Difficulty in IR Evaluation: The Effect of Systems, Corpora,
and System Components},
author = {Fabio Zampieri and Kevin Roitero and Shane J Culpepper and Oren Kurland and Stefano Mizzaro},
url = {https://doi.org/10.1145/3331184.3331279},
doi = {10.1145/3331184.3331279},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 42nd International ACM SIGIR Conference on
Research and Development in Information Retrieval, SIGIR 2019, Paris,
France, July 21-25, 2019.},
pages = {909--912},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Han, Lei; Roitero, Kevin; Gadiraju, Ujwal; Sarasua, Cristina; Checco, Alessandro; Maddalena, Eddy; Demartini, Gianluca
All Those Wasted Hours: On Task Abandonment in Crowdsourcing Proceedings Article
In: Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, WSDM 2019, Melbourne, VIC, Australia, February 11-15, 2019, pp. 321–329, 2019.
@inproceedings{DBLP:conf/wsdm/HanRGSCMD19,
title = {All Those Wasted Hours: On Task Abandonment in Crowdsourcing},
author = {Lei Han and Kevin Roitero and Ujwal Gadiraju and Cristina Sarasua and Alessandro Checco and Eddy Maddalena and Gianluca Demartini},
url = {https://doi.org/10.1145/3289600.3291035},
doi = {10.1145/3289600.3291035},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the Twelfth ACM International Conference on Web Search
and Data Mining, WSDM 2019, Melbourne, VIC, Australia, February
11-15, 2019},
pages = {321--329},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soprano, Michael; Mizzaro, Stefano
Crowdsourcing Peer Review: As We May Do Proceedings Article
In: Manghi, Paolo; Candela, Leonardo; Silvello, Gianmaria (Ed.): Digital Libraries: Supporting Open Science, pp. 259–273, Springer International Publishing, Cham, 2019, ISBN: 978-3-030-11226-4.
@inproceedings{10.1007/978-3-030-11226-4_21,
title = {Crowdsourcing Peer Review: As We May Do},
author = {Michael Soprano and Stefano Mizzaro},
editor = {Paolo Manghi and Leonardo Candela and Gianmaria Silvello},
isbn = {978-3-030-11226-4},
year = {2019},
date = {2019-01-01},
booktitle = {Digital Libraries: Supporting Open Science},
pages = {259--273},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {This paper describes Readersourcing 2.0, an ecosystem providing an implementation of the Readersourcing approach proposed by Mizzaro [10]. Readersourcing is proposed as an alternative to the standard peer review activity that aims to exploit the otherwise lost opinions of readers. Readersourcing 2.0 implements two different models based on the so-called codetermination algorithms. We describe the requirements, present the overall architecture, and show how the end-user can interact with the system. Readersourcing 2.0 will be used in the future to study also other topics, like the idea of shepherding the users to achieve a better quality of the reviews and the differences between a review activity carried out with a single-blind or a double-blind approach.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Roitero, Kevin; Brunello, Andrea; Urbano, Julián; Mizzaro, Stefano
Towards Stochastic Simulations of Relevance Profiles Proceedings Article
In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 2217–2220, Association for Computing Machinery, Beijing, China, 2019, ISBN: 9781450369763.
@inproceedings{10.1145/3357384.3358123,
title = {Towards Stochastic Simulations of Relevance Profiles},
author = {Kevin Roitero and Andrea Brunello and Julián Urbano and Stefano Mizzaro},
url = {https://doi.org/10.1145/3357384.3358123},
doi = {10.1145/3357384.3358123},
isbn = {9781450369763},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 28th ACM International Conference on Information and Knowledge Management},
pages = {2217–2220},
publisher = {Association for Computing Machinery},
address = {Beijing, China},
series = {CIKM '19},
abstract = {Recently proposed methods allow the generation of simulated scores representing the values of an effectiveness metric, but they do not investigate the generation of the actual lists of retrieved documents. In this paper we address this limitation: we present an approach that exploits an evolutionary algorithm and, given a metric score, creates a simulated relevance profile (i.e., a ranked list of relevance values) that produces that score. We show how the simulated relevance profiles are realistic under various analyses.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Han, Lei; Roitero, Kevin; Maddalena, Eddy; Mizzaro, Stefano; Demartini, Gianluca
On Transforming Relevance Scales Proceedings Article
In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 39–48, Association for Computing Machinery, Beijing, China, 2019, ISBN: 9781450369763.
@inproceedings{10.1145/3357384.3357988,
title = {On Transforming Relevance Scales},
author = {Lei Han and Kevin Roitero and Eddy Maddalena and Stefano Mizzaro and Gianluca Demartini},
url = {https://doi.org/10.1145/3357384.3357988},
doi = {10.1145/3357384.3357988},
isbn = {9781450369763},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 28th ACM International Conference on Information and Knowledge Management},
pages = {39–48},
publisher = {Association for Computing Machinery},
address = {Beijing, China},
series = {CIKM '19},
abstract = {Information Retrieval (IR) researchers have often used existing IR evaluation collections and transformed the relevance scale in which judgments have been collected, e.g., to use metrics that assume binary judgments like Mean Average Precision. Such scale transformations are often arbitrary (e.g., 0,1 mapped to 0 and 2,3 mapped to 1) and it is assumed that they have no impact on the results of IR evaluation. Moreover, the use of crowdsourcing to collect relevance judgments has become a standard methodology. When designing the crowdsourcing relevance judgment task, one of the decision to be made is the how granular the relevance scale used to collect judgments should be. Such decision has then repercussions on the metrics used to measure IR system effectiveness. In this paper we look at the effect of scale transformations in a systematic way. We perform extensive experiments to study the transformation of judgments from fine-grained to coarse-grained. We use different relevance judgments expressed on different relevance scales and either expressed by expert annotators or collected by means of crowdsourcing. The objective is to understand the impact of relevance scale transformations on IR evaluation outcomes and to draw conclusions on how to best transform judgments into a different scale, when necessary.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soprano, Kevin Roitero
Bias and Fairness in Effectiveness Evaluation by Means of Network Analysis and Mixture Models Proceedings Article
In: Agosti, Maristella; Buccio, Emanuele Di; Melucci, Massimo; Mizzaro, Stefano; Pasi, Gabriella; Silvestri, Fabrizio (Ed.): Proceedings of the 10th Italian Information Retrieval Workshop, Padova, Italy, September 16-18, 2019, pp. 6–7, CEUR-WS.org, 2019.
@inproceedings{workshop-paper-iir2019,
title = {Bias and Fairness in Effectiveness Evaluation by Means of Network Analysis and Mixture Models},
author = {Kevin Roitero Soprano},
editor = {Maristella Agosti and Emanuele Di Buccio and Massimo Melucci and Stefano Mizzaro and Gabriella Pasi and Fabrizio Silvestri},
url = {http://ceur-ws.org/Vol-2441/paper4.pdf},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 10th Italian Information Retrieval Workshop, Padova, Italy, September 16-18, 2019},
volume = {2441},
pages = {6--7},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soprano, Michael; Roitero, Kevin; Mizzaro, Stefano
HITS Hits Readersourcing: Validating Peer Review Alternatives Using Network Analysis Proceedings Article
In: Chandrasekaran, Muthu Kumar; Mayr, Philipp (Ed.): Proceedings of the 4th Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries (BIRNDL 2019) co-located with the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2019), Paris, France, July 25, 2019, pp. 70–82, CEUR-WS.org, 2019.
@inproceedings{workshop-paper-birnld2019,
title = {HITS Hits Readersourcing: Validating Peer Review Alternatives Using Network Analysis},
author = {Michael Soprano and Kevin Roitero and Stefano Mizzaro},
editor = {Muthu Kumar Chandrasekaran and Philipp Mayr},
url = {http://ceur-ws.org/Vol-2414/paper7.pdf},
year = {2019},
date = {2019-01-01},
booktitle = {Proceedings of the 4th Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries (BIRNDL 2019) co-located with the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2019), Paris, France, July 25, 2019},
volume = {2414},
pages = {70--82},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soprano, Michael; Mizzaro, Stefano
Crowdsourcing Peer Review: As We May Do Proceedings Article
In: Manghi, Paolo; Candela, Leonardo; Silvello, Gianmaria (Ed.): Digital Libraries: Supporting Open Science - 15th Italian Research Conference on Digital Libraries, IRCDL 2019, Pisa, Italy, January 31 - February 1, 2019, Proceedings, pp. 259–273, Springer, 2019.
@inproceedings{conference-paper-ircdl2019,
title = {Crowdsourcing Peer Review: As We May Do},
author = {Michael Soprano and Stefano Mizzaro},
editor = {Paolo Manghi and Leonardo Candela and Gianmaria Silvello},
url = {https://doi.org/10.1007/978-3-030-11226-4_21},
doi = {10.1007/978-3-030-11226-4_21},
year = {2019},
date = {2019-01-01},
booktitle = {Digital Libraries: Supporting Open Science - 15th Italian Research Conference on Digital Libraries, IRCDL 2019, Pisa, Italy, January 31 - February 1, 2019, Proceedings},
volume = {988},
pages = {259--273},
publisher = {Springer},
series = {Communications in Computer and Information Science},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soprano, Michael; Mizzaro, Stefano
Crowdsourcing Peer Review in the Digital Humanities? Proceedings Article
In: Book of Abstracts, 8th AIUCD Conference 2019 – Pedagogy, Teaching, and Research in the Age of Digital Humanities, page 251. Udine, Italy, 2019, pp. 251, Udine, Italy, 2019, ISBN: 9788894253535.
@inproceedings{conference-paper-aiudc2019,
title = {Crowdsourcing Peer Review in the Digital Humanities?},
author = {Michael Soprano and Stefano Mizzaro},
url = {http://aiucd2019.uniud.it/wp-content/uploads/2020/03/AIUCD2019-BoA_DEF.pdf},
isbn = {9788894253535},
year = {2019},
date = {2019-01-01},
booktitle = {Book of Abstracts, 8th AIUCD Conference 2019 – Pedagogy, Teaching, and Research in the Age of Digital Humanities, page 251. Udine, Italy, 2019},
pages = {251},
address = {Udine, Italy},
series = {AIUCD '19},
abstract = {We propose an alternative approach to the standard peer review activity that aims to exploit the otherwise lost opinions of readers of publications which is called Readersourcing, originally proposed by Mizzaro [1]. Such an approach can be formalized by means of different models which share the same general principles. These models should be able to define a way, to measure the overall quality of a publication as well the reputation of a reader as an assessor; moreover, from these measures it should be possible to derive the reputation of a scholar as an author. We describe an ecosystem called Readersourcing 2.0 which provides an implementation for two Readersourcing models [2, 3] by outlining its goals and requirements. Readersourcing 2.0 will be used in the future to gather fresh data to analyze and validate.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2018
Roitero, Kevin; Spina, Damiano; Demartini, Gianluca; Mizzaro, Stefano
How Many Truth Levels? Six? One Hundred? Even more? Validating Truthfulness of Statements via Crowdsourcing Proceedings Article
In: To appear in RDSM 2018 : Workshop on Rumours and Deception in Social Media., 2018.
@inproceedings{toapp:howmany,
title = {How Many Truth Levels? Six? One Hundred? Even more? Validating Truthfulness of Statements via Crowdsourcing},
author = {Kevin Roitero and Damiano Spina and Gianluca Demartini and Stefano Mizzaro},
year = {2018},
date = {2018-01-01},
booktitle = {To appear in RDSM 2018 : Workshop on Rumours and Deception in Social Media.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Roitero, Kevin; Soprano, Michael; Mizzaro, Stefano
Effectiveness Evaluation with a Subset of Topics: A Practical Approach Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 1145–1148, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210108,
title = {Effectiveness Evaluation with a Subset of Topics: A Practical Approach},
author = {Kevin Roitero and Michael Soprano and Stefano Mizzaro},
url = {https://doi.org/10.1145/3209978.3210108},
doi = {10.1145/3209978.3210108},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {1145–1148},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {Several researchers have proposed to reduce the number of topics used in TREC-like initiatives. One research direction that has been pursued is what is the optimal topic subset of a given cardinality that evaluates the systems/runs in the most accurate way. Such a research direction has been so far mainly theoretical, with almost no indication on how to select the few good topics in practice. We propose such a practical criterion for topic selection: we rely on the methods for automatic system evaluation without relevance judgments, and by running some experiments on several TREC collections we show that the topics selected on the basis of those evaluations are indeed more informative than random topics.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Roitero, Kevin; Maddalena, Eddy; Ponte, Yannick; Mizzaro, Stefano
IRevalOO: An Object Oriented Framework for Retrieval Evaluation Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 913–916, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210084,
title = {IRevalOO: An Object Oriented Framework for Retrieval Evaluation},
author = {Kevin Roitero and Eddy Maddalena and Yannick Ponte and Stefano Mizzaro},
url = {https://doi.org/10.1145/3209978.3210084},
doi = {10.1145/3209978.3210084},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {913–916},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {We propose IRevalOO, a flexible Object Oriented framework that (i) can be used as-is as a replacement of the widely adopted trec_eval software, and (ii) can be easily extended (or "instantiated'', in framework terminology) to implement different scenarios of test collection based retrieval evaluation. Instances of IRevalOO can provide a usable and convenient alternative to the state-of-the-art software commonly used by different initiatives (TREC, NTCIR, CLEF, FIRE, etc.). Also, those instances can be easily adapted to satisfy future customization needs of researchers, as: implementing and experimenting with new metrics, even based on new notions of relevance; using different formats for system output and "qrels''; and in general visualizing, comparing, and managing retrieval evaluation results.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Roitero, Kevin
CHEERS: CHeap & Engineered Evaluation of Retrieval Systems Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 1467, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210229,
title = {CHEERS: CHeap & Engineered Evaluation of Retrieval Systems},
author = {Kevin Roitero},
url = {https://doi.org/10.1145/3209978.3210229},
doi = {10.1145/3209978.3210229},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {1467},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {In test collection based evaluation of retrieval effectiveness, many research investigated different directions for an economical and a semi-automatic evaluation of retrieval systems. Although several methods have been proposed and experimentally evaluated, their accuracy seems still limited. In this paper we present our proposal for a more engineered approach to information retrieval evaluation.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Roitero, Kevin; Maddalena, Eddy; Demartini, Gianluca; Mizzaro, Stefano
On Fine-Grained Relevance Scales Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 675–684, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210052,
title = {On Fine-Grained Relevance Scales},
author = {Kevin Roitero and Eddy Maddalena and Gianluca Demartini and Stefano Mizzaro},
url = {https://doi.org/10.1145/3209978.3210052},
doi = {10.1145/3209978.3210052},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
urldate = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {675–684},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {In Information Retrieval evaluation, the classical approach of adopting binary relevance judgments has been replaced by multi-level relevance judgments and by gain-based metrics leveraging such multi-level judgment scales. Recent work has also proposed and evaluated unbounded relevance scales by means of Magnitude Estimation (ME) and compared them with multi-level scales. While ME brings advantages like the ability for assessors to always judge the next document as having higher or lower relevance than any of the documents they have judged so far, it also comes with some drawbacks. For example, it is not a natural approach for human assessors to judge items as they are used to do on the Web (e.g., 5-star rating). In this work, we propose and experimentally evaluate a bounded and fine-grained relevance scale having many of the advantages and dealing with some of the issues of ME. We collect relevance judgments over a 100-level relevance scale (S100) by means of a large-scale crowdsourcing experiment and compare the results with other relevance scales (binary, 4-level, and ME) showing the benefit of fine-grained scales over both coarse-grained and unbounded scales as well as highlighting some new results on ME. Our results show that S100 maintains the flexibility of unbounded scales like ME in providing assessors with ample choice when judging document relevance (i.e., assessors can fit relevance judgments in between of previously given judgments). It also allows assessors to judge on a more familiar scale (e.g., on 10 levels) and to perform efficiently since the very first judging task.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Mizzaro, Stefano; Mothe, Josiane; Roitero, Kevin; Ullah, Md Zia
Query Performance Prediction and Effectiveness Evaluation Without Relevance Judgments: Two Sides of the Same Coin Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 1233–1236, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210146,
title = {Query Performance Prediction and Effectiveness Evaluation Without Relevance Judgments: Two Sides of the Same Coin},
author = {Stefano Mizzaro and Josiane Mothe and Kevin Roitero and Md Zia Ullah},
url = {https://doi.org/10.1145/3209978.3210146},
doi = {10.1145/3209978.3210146},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {1233–1236},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {Some methods have been developed for automatic effectiveness evaluation without relevance judgments. We propose to use those methods, and their combination based on a machine learning approach, for query performance prediction. Moreover, since predicting average precision as it is usually done in query performance prediction literature is sensitive to the reference system that is chosen, we focus on predicting the average of average precision values over several systems. Results of an extensive experimental evaluation on ten TREC collections show that our proposed methods outperform state-of-the-art query performance predictors.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Giner, Fernando; Mizzaro, Stefano; Spina, Damiano
A Formal Account of Effectiveness Evaluation and Ranking Fusion Proceedings Article
In: Proceedings of the 2018 ACM SIGIR International Conference on Theory of Information Retrieval, pp. 123–130, Association for Computing Machinery, Tianjin, China, 2018, ISBN: 9781450356565.
@inproceedings{10.1145/3234944.3234958,
title = {A Formal Account of Effectiveness Evaluation and Ranking Fusion},
author = {Enrique Amigó and Fernando Giner and Stefano Mizzaro and Damiano Spina},
url = {https://doi.org/10.1145/3234944.3234958},
doi = {10.1145/3234944.3234958},
isbn = {9781450356565},
year = {2018},
date = {2018-01-01},
booktitle = {Proceedings of the 2018 ACM SIGIR International Conference on Theory of Information Retrieval},
pages = {123–130},
publisher = {Association for Computing Machinery},
address = {Tianjin, China},
series = {ICTIR '18},
abstract = {This paper proposes a theoretical framework which models the information provided by retrieval systems in terms of Information Theory. The proposed framework allows to formalize: (i) system effectiveness as an information theoretic similarity between system outputs and human assessments, and (ii) ranking fusion as an information quantity measure. As a result, the proposed effectiveness metric improves popular metrics in terms of formal constraints. In addition, our empirical experiments suggest that it captures quality aspects from traditional metrics, while the reverse is not true. Our work also advances the understanding of theoretical foundations of the empirically known phenomenon of effectiveness increase when combining retrieval system outputs in an unsupervised manner.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Fang, Hui; Mizzaro, Stefano; Zhai, ChengXiang
Are We on the Right Track? An Examination of Information Retrieval Methodologies Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 997–1000, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210131b,
title = {Are We on the Right Track? An Examination of Information Retrieval Methodologies},
author = {Enrique Amigó and Hui Fang and Stefano Mizzaro and ChengXiang Zhai},
url = {https://doi.org/10.1145/3209978.3210131},
doi = {10.1145/3209978.3210131},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {997–1000},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {The unpredictability of user behavior and the need for effectiveness make it difficult to define a suitable research methodology for Information Retrieval (IR). In order to tackle this challenge, we categorize existing IR methodologies along two dimensions: (1) empirical vs. theoretical, and (2) top-down vs. bottom-up. The strengths and drawbacks of the resulting categories are characterized according to 6 desirable aspects. The analysis suggests that different methodologies are complementary and therefore, equally necessary. The categorization of the 167 full papers published in the last SIGIR (2016 and 2017) and ICTIR (2017) conferences suggest that most of existing work is empirical bottom-up, suggesting lack of some desirable aspects. With the hope of improving IR research practice, we propose a general methodology for IR that integrates the strengths of existing research methods.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Fang, Hui; Mizzaro, Stefano; Zhai, ChengXiang
Report on the SIGIR 2017 Workshop on Axiomatic Thinking for Information Retrieval and Related Tasks (ATIR) Journal Article
In: SIGIR Forum, vol. 51, no 3, pp. 99–106, 2018, ISSN: 0163-5840.
@article{10.1145/3190580.3190596,
title = {Report on the SIGIR 2017 Workshop on Axiomatic Thinking for Information Retrieval and Related Tasks (ATIR)},
author = {Enrique Amigó and Hui Fang and Stefano Mizzaro and ChengXiang Zhai},
url = {https://doi.org/10.1145/3190580.3190596},
doi = {10.1145/3190580.3190596},
issn = {0163-5840},
year = {2018},
date = {2018-01-01},
journal = {SIGIR Forum},
volume = {51},
number = {3},
pages = {99–106},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
abstract = {The SIGIR 2017 workshop on Axiomatic Thinking for Information Retrieval and Related Tasks took place on August 11, 2017 in Tokyo, Japan. The workshop aimed to help foster collaboration of researchers working on different perspectives of axiomatic thinking and encourage discussion and research on general methodological issues related to applying axiomatic thinking to information retrieval and related tasks. The program consisted of one keynote talk, four research presentations and a final panel discussion. This report outlines the events of the workshop and summarizes the major outcomes. More information about the workshop is available at https://www.eecis.udel.edu/~hfang/ATIR.html.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Roitero, Kevin; Passon, Marco; Serra, Giuseppe; Mizzaro, Stefano
Reproduce. Generalize. Extend. On Information Retrieval Evaluation without Relevance Judgments Journal Article
In: J. Data and Information Quality, vol. 10, no 3, 2018, ISSN: 1936-1955.
@article{10.1145/3241064,
title = {Reproduce. Generalize. Extend. On Information Retrieval Evaluation without Relevance Judgments},
author = {Kevin Roitero and Marco Passon and Giuseppe Serra and Stefano Mizzaro},
url = {https://doi.org/10.1145/3241064},
doi = {10.1145/3241064},
issn = {1936-1955},
year = {2018},
date = {2018-01-01},
journal = {J. Data and Information Quality},
volume = {10},
number = {3},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
abstract = {The evaluation of retrieval effectiveness by means of test collections is a commonly used methodology in the information retrieval field. Some researchers have addressed the quite fascinating research question of whether it is possible to evaluate effectiveness completely automatically, without human relevance assessments. Since human relevance assessment is one of the main costs of building a test collection, both in human time and money resources, this rather ambitious goal would have a practical impact. In this article, we reproduce the main results on evaluating information retrieval systems without relevance judgments; furthermore, we generalize such previous work to analyze the effect of test collections, evaluation metrics, and pool depth. We also expand the idea to semi-automatic evaluation and estimation of topic difficulty. Our results show that (i) previous work is overall reproducible, although some specific results are not; (ii) collection, metric, and pool depth impact the automatic evaluation of systems, which is anyway accurate in several cases; (iii) semi-automatic evaluation is an effective methodology; and (iv) automatic evaluation can (to some extent) be used to predict topic difficulty.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Roitero, Kevin; Soprano, Michael; Mizzaro, Stefano
Effectiveness Evaluation with a Subset of Topics: A Practical Approach Proceedings Article
In: Proceedings of the 41st International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2018). Ann Arbor Michigan, U.S.A, July 8-12, 2018. Conference Rank: GGS A++, Core A*, pp. 1145–1148, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{conference-paper-sigir2018,
title = {Effectiveness Evaluation with a Subset of Topics: A Practical Approach},
author = {Kevin Roitero and Michael Soprano and Stefano Mizzaro},
url = {https://doi.org/10.1145/3209978.3210108},
doi = {10.1145/3209978.3210108},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {Proceedings of the 41st International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2018). Ann Arbor Michigan, U.S.A, July 8-12, 2018. Conference Rank: GGS A++, Core A*},
pages = {1145–1148},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {Several researchers have proposed to reduce the number of topics used in TREC-like initiatives. One research direction that has been pursued is what is the optimal topic subset of a given cardinality that evaluates the systems/runs in the most accurate way. Such a research direction has been so far mainly theoretical, with almost no indication on how to select the few good topics in practice. We propose such a practical criterion for topic selection: we rely on the methods for automatic system evaluation without relevance judgments, and by running some experiments on several TREC collections we show that the topics selected on the basis of those evaluations are indeed more informative than random topics.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Maddalena, Eddy; Ceolin, Davide; Mizzaro, Stefano
Multidimensional News Quality: A Comparison of Crowdsourcing and Nichesourcing Proceedings Article
In: Cuzzocrea, Alfredo; Bonchi, Francesco; Gunopulos, Dimitrios (Ed.): Proceedings of the CIKM 2018 Workshops co-located with 27th ACM International Conference on Information and Knowledge Management (CIKM 2018), Torino, Italy, October 22, 2018, CEUR-WS.org, 2018.
@inproceedings{DBLP:conf/cikm/MaddalenaCM18,
title = {Multidimensional News Quality: A Comparison of Crowdsourcing and
Nichesourcing},
author = {Eddy Maddalena and Davide Ceolin and Stefano Mizzaro},
editor = {Alfredo Cuzzocrea and Francesco Bonchi and Dimitrios Gunopulos},
url = {http://ceur-ws.org/Vol-2482/paper17.pdf},
year = {2018},
date = {2018-01-01},
booktitle = {Proceedings of the CIKM 2018 Workshops co-located with 27th ACM
International Conference on Information and Knowledge Management (CIKM
2018), Torino, Italy, October 22, 2018},
volume = {2482},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Fang, Hui; Mizzaro, Stefano; Zhai, ChengXiang
Are we on the Right Track? An Examination of Information Retrieval Methodologies Proceedings Article
In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 997–1000, Association for Computing Machinery, Ann Arbor, MI, USA, 2018, ISBN: 9781450356572.
@inproceedings{10.1145/3209978.3210131,
title = {Are we on the Right Track? An Examination of Information Retrieval Methodologies},
author = {Enrique Amigó and Hui Fang and Stefano Mizzaro and ChengXiang Zhai},
url = {https://doi.org/10.1145/3209978.3210131},
doi = {10.1145/3209978.3210131},
isbn = {9781450356572},
year = {2018},
date = {2018-01-01},
booktitle = {The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {997–1000},
publisher = {Association for Computing Machinery},
address = {Ann Arbor, MI, USA},
series = {SIGIR '18},
abstract = {The unpredictability of user behavior and the need for effectiveness make it difficult to define a suitable research methodology for Information Retrieval (IR). In order to tackle this challenge, we categorize existing IR methodologies along two dimensions: (1) empirical vs. theoretical, and (2) top-down vs. bottom-up. The strengths and drawbacks of the resulting categories are characterized according to 6 desirable aspects. The analysis suggests that different methodologies are complementary and therefore, equally necessary. The categorization of the 167 full papers published in the last SIGIR (2016 and 2017) and ICTIR (2017) conferences suggest that most of existing work is empirical bottom-up, suggesting lack of some desirable aspects. With the hope of improving IR research practice, we propose a general methodology for IR that integrates the strengths of existing research methods.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Giner, Fernando; Mizzaro, Stefano; Spina, Damiano
A Formal Account of Effectiveness Evaluation and Ranking Fusion Proceedings Article
In: Proceedings of the 2018 ACM SIGIR International Conference on Theory of Information Retrieval, pp. 123–130, Association for Computing Machinery, Tianjin, China, 2018, ISBN: 9781450356565.
@inproceedings{10.1145/3234944.3234958b,
title = {A Formal Account of Effectiveness Evaluation and Ranking Fusion},
author = {Enrique Amigó and Fernando Giner and Stefano Mizzaro and Damiano Spina},
url = {https://doi.org/10.1145/3234944.3234958},
doi = {10.1145/3234944.3234958},
isbn = {9781450356565},
year = {2018},
date = {2018-01-01},
booktitle = {Proceedings of the 2018 ACM SIGIR International Conference on Theory of Information Retrieval},
pages = {123–130},
publisher = {Association for Computing Machinery},
address = {Tianjin, China},
series = {ICTIR '18},
abstract = {This paper proposes a theoretical framework which models the information provided by retrieval systems in terms of Information Theory. The proposed framework allows to formalize: (i) system effectiveness as an information theoretic similarity between system outputs and human assessments, and (ii) ranking fusion as an information quantity measure. As a result, the proposed effectiveness metric improves popular metrics in terms of formal constraints. In addition, our empirical experiments suggest that it captures quality aspects from traditional metrics, while the reverse is not true. Our work also advances the understanding of theoretical foundations of the empirically known phenomenon of effectiveness increase when combining retrieval system outputs in an unsupervised manner.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Fang, Hui; Mizzaro, Stefano; Zhai, ChengXiang
Report on the SIGIR 2017 Workshop on Axiomatic Thinking for Information Retrieval and Related Tasks (ATIR) Journal Article
In: SIGIR Forum, vol. 51, no 3, pp. 99–106, 2018, ISSN: 0163-5840.
@article{10.1145/3190580.3190596b,
title = {Report on the SIGIR 2017 Workshop on Axiomatic Thinking for Information Retrieval and Related Tasks (ATIR)},
author = {Enrique Amigó and Hui Fang and Stefano Mizzaro and ChengXiang Zhai},
url = {https://doi.org/10.1145/3190580.3190596},
doi = {10.1145/3190580.3190596},
issn = {0163-5840},
year = {2018},
date = {2018-01-01},
journal = {SIGIR Forum},
volume = {51},
number = {3},
pages = {99–106},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
abstract = {The SIGIR 2017 workshop on Axiomatic Thinking for Information Retrieval and Related Tasks took place on August 11, 2017 in Tokyo, Japan. The workshop aimed to help foster collaboration of researchers working on different perspectives of axiomatic thinking and encourage discussion and research on general methodological issues related to applying axiomatic thinking to information retrieval and related tasks. The program consisted of one keynote talk, four research presentations and a final panel discussion. This report outlines the events of the workshop and summarizes the major outcomes. More information about the workshop is available at https://www.eecis.udel.edu/~hfang/ATIR.html.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2017
Checco, Alessandro; Roitero, Kevin; Maddalena, Eddy; Mizzaro, Stefano; Demartini, Gianluca
Let’s Agree to Disagree: Fixing Agreement Measures for Crowdsourcing Journal Article
In: Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, vol. 5, no 1, 2017.
@article{Checco_Roitero_Maddalena_Mizzaro_Demartini_2017,
title = {Let’s Agree to Disagree: Fixing Agreement Measures for Crowdsourcing},
author = {Alessandro Checco and Kevin Roitero and Eddy Maddalena and Stefano Mizzaro and Gianluca Demartini},
url = {https://ojs.aaai.org/index.php/HCOMP/article/view/13306},
year = {2017},
date = {2017-09-01},
journal = {Proceedings of the AAAI Conference on Human Computation and Crowdsourcing},
volume = {5},
number = {1},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Roitero, Kevin; Mizzaro, Stefano; Serra, Giuseppe
Economic Evaluation of Recommender Systems: A Proposal Proceedings Article
In: Proceedings of the 8th Italian Information Retrieval Workshop, Lugano, Switzerland, June 05-07, 2017., pp. 48–51, 2017.
@inproceedings{DBLP:conf/iir/RoiteroMS17,
title = {Economic Evaluation of Recommender Systems: A Proposal},
author = {Kevin Roitero and Stefano Mizzaro and Giuseppe Serra},
url = {http://ceur-ws.org/Vol-1911/8.pdf},
year = {2017},
date = {2017-01-01},
booktitle = {Proceedings of the 8th Italian Information Retrieval Workshop, Lugano,
Switzerland, June 05-07, 2017.},
pages = {48--51},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Maddalena, Eddy; Roitero, Kevin; Demartini, Gianluca; Mizzaro, Stefano
Considering Assessor Agreement in IR Evaluation Proceedings Article
In: Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval, pp. 75–82, Association for Computing Machinery, New York, NY, USA, 2017, ISBN: 9781450344906.
@inproceedings{10.1145/3121050.3121060b,
title = {Considering Assessor Agreement in IR Evaluation},
author = {Eddy Maddalena and Kevin Roitero and Gianluca Demartini and Stefano Mizzaro},
url = {https://doi.org/10.1145/3121050.3121060},
doi = {10.1145/3121050.3121060},
isbn = {9781450344906},
year = {2017},
date = {2017-01-01},
booktitle = {Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval},
pages = {75–82},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
abstract = {The agreement between relevance assessors is an important but understudied topic in the Information Retrieval literature because of the limited data available about documents assessed by multiple judges. This issue has gained even more importance recently in light of crowdsourced relevance judgments, where it is customary to gather many relevance labels for each topic-document pair. In a crowdsourcing setting, agreement is often even used as a proxy for quality, although without any systematic verification of the conjecture that higher agreement corresponds to higher quality. In this paper we address this issue and we study in particular: the effect of topic on assessor agreement; the relationship between assessor agreement and judgment quality; the effect of agreement on ranking systems according to their effectiveness; and the definition of an agreement-aware effectiveness metric that does not discard information about multiple judgments for the same document as it typically happens in a crowdsourcing setting.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Chifu, Adrian-Gabriel; Déjean, Sébastien; Mizzaro, Stefano; Mothe, Josiane
Human-Based Query Difficulty Prediction Proceedings Article
In: 39th European Colloquium on Information Retrieval (ECIR 2017), pp. pp. 343-356, Aberdeen, Scotland, United Kingdom, 2017.
@inproceedings{chifu:hal-01712541,
title = {Human-Based Query Difficulty Prediction},
author = {Adrian-Gabriel Chifu and Sébastien Déjean and Stefano Mizzaro and Josiane Mothe},
url = {https://hal.archives-ouvertes.fr/hal-01712541},
year = {2017},
date = {2017-01-01},
booktitle = {39th European Colloquium on Information Retrieval (ECIR 2017)},
pages = {pp. 343-356},
address = {Aberdeen, Scotland, United Kingdom},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Maddalena, Eddy; Mizzaro, Stefano; Scholer, Falk; Turpin, Andrew
On Crowdsourcing Relevance Magnitudes for Information Retrieval Evaluation Journal Article
In: ACM Trans. Inf. Syst., vol. 35, no 3, 2017, ISSN: 1046-8188.
@article{10.1145/3002172,
title = {On Crowdsourcing Relevance Magnitudes for Information Retrieval Evaluation},
author = {Eddy Maddalena and Stefano Mizzaro and Falk Scholer and Andrew Turpin},
url = {https://doi.org/10.1145/3002172},
doi = {10.1145/3002172},
issn = {1046-8188},
year = {2017},
date = {2017-01-01},
journal = {ACM Trans. Inf. Syst.},
volume = {35},
number = {3},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
abstract = {Magnitude estimation is a psychophysical scaling technique for the measurement of sensation, where observers assign numbers to stimuli in response to their perceived intensity. We investigate the use of magnitude estimation for judging the relevance of documents for information retrieval evaluation, carrying out a large-scale user study across 18 TREC topics and collecting over 50,000 magnitude estimation judgments using crowdsourcing. Our analysis shows that magnitude estimation judgments can be reliably collected using crowdsourcing, are competitive in terms of assessor cost, and are, on average, rank-aligned with ordinal judgments made by expert relevance assessors.We explore the application of magnitude estimation for IR evaluation, calibrating two gain-based effectiveness metrics, nDCG and ERR, directly from user-reported perceptions of relevance. A comparison of TREC system effectiveness rankings based on binary, ordinal, and magnitude estimation relevance shows substantial variation; in particular, the top systems ranked using magnitude estimation and ordinal judgments differ substantially. Analysis of the magnitude estimation scores shows that this effect is due in part to varying perceptions of relevance: different users have different perceptions of the impact of relative differences in document relevance. These results have direct implications for IR evaluation, suggesting that current assumptions about a single view of relevance being sufficient to represent a population of users are unlikely to hold.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Crestani, Fabio; Mizzaro, Stefano; Scagnetto, Ivan
MobileInformation Retrieval Book
1, 2017, ISBN: 978-3-319-60777-1.
@book{mir-miz-cre-sca,
title = {MobileInformation Retrieval},
author = {Fabio Crestani and Stefano Mizzaro and Ivan Scagnetto},
editor = {Springer. Cham},
doi = {10.1007/978-3-319-60777-1},
isbn = {978-3-319-60777-1},
year = {2017},
date = {2017-01-01},
edition = {1},
keywords = {},
pubstate = {published},
tppubtype = {book}
}
Pavan, Marco; Mizzaro, Stefano; Scagnetto, Ivan
Mining Movement Data to Extract Personal Points of Interest: A Feature Based Approach Book Chapter
In: Lai, Cristian; Giuliani, Alessandro; Semeraro, Giovanni (Ed.): Information Filtering and Retrieval: DART 2014: Revised and Invited Papers, pp. 35–61, Springer International Publishing, Cham, 2017, ISBN: 978-3-319-46135-9.
@inbook{Pavan2017,
title = {Mining Movement Data to Extract Personal Points of Interest: A Feature Based Approach},
author = {Marco Pavan and Stefano Mizzaro and Ivan Scagnetto},
editor = {Cristian Lai and Alessandro Giuliani and Giovanni Semeraro},
url = {https://doi.org/10.1007/978-3-319-46135-9_3},
doi = {10.1007/978-3-319-46135-9_3},
isbn = {978-3-319-46135-9},
year = {2017},
date = {2017-01-01},
booktitle = {Information Filtering and Retrieval: DART 2014: Revised and Invited Papers},
pages = {35--61},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {Due to the widespread of mobile devices in recent years, records of the locations visited by users are common and growing, and the availability of such large amounts of spatio-temporal data opens new challenges to automatically discover valuable knowledge. One aspect that is being studied is the identification of important locations, i.e. places where people spend a fair amount of time during their daily activities; we address it with a novel approach. Our proposed method is organised in two phases: first, a set of candidate stay points is identified by exploiting some state-of-the-art algorithms to filter the GPS-logs; then, the candidate stay points are mapped onto a feature space having as dimensions the area underlying the stay point, its intensity (e.g. the time spent in a location) and its frequency (e.g. the number of total visits). We conjecture that the feature space allows to model aspects/measures that are more semantically related to users and better suited to reason about their similarities and differences than simpler physical measures (e.g. latitude, longitude, and timestamp). An experimental evaluation on the GeoLife public dataset confirms the effectiveness of our approach and sheds some light on the peculiar features and critical issues of location based systems.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Amigó, Enrique; Fang, Hui; Mizzaro, Stefano; Zhai, ChengXiang
Axiomatic Thinking for Information Retrieval: And Related Tasks Proceedings Article
In: Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1419–1420, Association for Computing Machinery, Shinjuku, Tokyo, Japan, 2017, ISBN: 9781450350228.
@inproceedings{10.1145/3077136.3084369,
title = {Axiomatic Thinking for Information Retrieval: And Related Tasks},
author = {Enrique Amigó and Hui Fang and Stefano Mizzaro and ChengXiang Zhai},
url = {https://doi.org/10.1145/3077136.3084369},
doi = {10.1145/3077136.3084369},
isbn = {9781450350228},
year = {2017},
date = {2017-01-01},
booktitle = {Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval},
pages = {1419–1420},
publisher = {Association for Computing Machinery},
address = {Shinjuku, Tokyo, Japan},
series = {SIGIR '17},
abstract = {This is the first workshop on the emerging interdisciplinary research area of applying axiomatic thinking to information retrieval (IR) and related tasks. The workshop aims to help foster collaboration of researchers working on different perspectives of axiomatic thinking and encourage discussion and research on general methodological issues related to applying axiomatic thinking to IR and related tasks.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Mea, Vincenzo Della; Baroni, Giulia L.; Pilutti, David; Loreto, Carla Di
SlideJ: An ImageJ plugin for automated processing of whole slide images Journal Article
In: PLOS ONE, vol. 12, no 7, pp. 1–9, 2017.
@article{10.1371/journal.pone.0180540,
title = {SlideJ: An ImageJ plugin for automated processing of whole slide images},
author = {Vincenzo Della Mea and Giulia L. Baroni and David Pilutti and Carla Di Loreto},
url = {https://doi.org/10.1371/journal.pone.0180540},
doi = {10.1371/journal.pone.0180540},
year = {2017},
date = {2017-01-01},
urldate = {2017-01-01},
journal = {PLOS ONE},
volume = {12},
number = {7},
pages = {1–9},
publisher = {Public Library of Science},
abstract = {The digital slide, or Whole Slide Image, is a digital image, acquired with specific scanners, that represents a complete tissue sample or cytological specimen at microscopic level. While Whole Slide image analysis is recognized among the most interesting opportunities, the typical size of such images—up to Gpixels- can be very demanding in terms of memory requirements. Thus, while algorithms and tools for processing and analysis of single microscopic field images are available, Whole Slide images size makes the direct use of such tools prohibitive or impossible. In this work a plugin for ImageJ, named SlideJ, is proposed with the objective to seamlessly extend the application of image analysis algorithms implemented in ImageJ for single microscopic field images to a whole digital slide analysis. The plugin has been complemented by examples of macro in the ImageJ scripting language to demonstrate its use in concrete situations.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2016
Roitero, Kevin; Mizzaro, Stefano
Improving the Efficiency of Retrieval Effectiveness Evaluation: Finding a Few Good Topics with Clustering? Proceedings Article
In: Proceedings of the 7th Italian Information Retrieval Workshop, Venezia, Italy, May 30-31, 2016., 2016.
@inproceedings{DBLP:conf/iir/RoiteroM16,
title = {Improving the Efficiency of Retrieval Effectiveness Evaluation: Finding
a Few Good Topics with Clustering?},
author = {Kevin Roitero and Stefano Mizzaro},
url = {http://ceur-ws.org/Vol-1653/paper_4.pdf},
year = {2016},
date = {2016-01-01},
booktitle = {Proceedings of the 7th Italian Information Retrieval Workshop, Venezia,
Italy, May 30-31, 2016.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Pavan, Marco; Mizzaro, Stefano; Bernardon, Matteo; Scagnetto, Ivan
Exploiting News to Categorize Tweets: Quantifying the Impact of Different News Collections Proceedings Article
In: -, Miguel Martinez; Kruschwitz, Udo; Kazai, Gabriella; Hopfgartner, Frank; Corney, David; Campos, Ricardo; Albakour, Dyaa (Ed.): Proceedings of the First International Workshop on Recent Trends in News Information Retrieval co-located with 38th European Conference on Information Retrieval (ECIR 2016), Padua, Italy, March 20, 2016, pp. 54–59, CEUR-WS.org, 2016.
@inproceedings{DBLP:conf/ecir/PavanMBS16,
title = {Exploiting News to Categorize Tweets: Quantifying the Impact of Different
News Collections},
author = {Marco Pavan and Stefano Mizzaro and Matteo Bernardon and Ivan Scagnetto},
editor = {Miguel Martinez - and Udo Kruschwitz and Gabriella Kazai and Frank Hopfgartner and David Corney and Ricardo Campos and Dyaa Albakour},
url = {http://ceur-ws.org/Vol-1568/paper10.pdf},
year = {2016},
date = {2016-01-01},
booktitle = {Proceedings of the First International Workshop on Recent Trends in
News Information Retrieval co-located with 38th European Conference
on Information Retrieval (ECIR 2016), Padua, Italy, March 20, 2016},
volume = {1568},
pages = {54--59},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2015
Pavan, M; Mizzaro, S; Scagnetto, I; Beggiato, A
Finding Important Locations: A Feature-Based Approach Proceedings Article
In: 2015 16th IEEE International Conference on Mobile Data Management, pp. 110-115, 2015.
@inproceedings{7264310b,
title = {Finding Important Locations: A Feature-Based Approach},
author = {M Pavan and S Mizzaro and I Scagnetto and A Beggiato},
doi = {10.1109/MDM.2015.11},
year = {2015},
date = {2015-01-01},
booktitle = {2015 16th IEEE International Conference on Mobile Data Management},
volume = {1},
pages = {110-115},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Mizzaro, Stefano; Pavan, Marco; Scagnetto, Ivan
Content-Based Similarity of Twitter Users Proceedings Article
In: Hanbury, Allan; Kazai, Gabriella; Rauber, Andreas; Fuhr, Norbert (Ed.): Advances in Information Retrieval, pp. 507–512, Springer International Publishing, Cham, 2015, ISBN: 978-3-319-16354-3.
@inproceedings{10.1007/978-3-319-16354-3_56,
title = {Content-Based Similarity of Twitter Users},
author = {Stefano Mizzaro and Marco Pavan and Ivan Scagnetto},
editor = {Allan Hanbury and Gabriella Kazai and Andreas Rauber and Norbert Fuhr},
isbn = {978-3-319-16354-3},
year = {2015},
date = {2015-01-01},
booktitle = {Advances in Information Retrieval},
pages = {507--512},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {We propose a method for computing user similarity based on a network representing the semantic relationships between the words occurring in the same tweet and the related topics. We use such specially crafted network to define several user profiles to be compared with cosine similarity. We also describe an initial experimental activity to study the effectiveness on a limited dataset.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amigó, Enrique; Gonzalo, Julio; Mizzaro, Stefano
A Formal Approach to Effectiveness Metrics for Information Access: Retrieval, Filtering, and Clustering Proceedings Article
In: Hanbury, Allan; Kazai, Gabriella; Rauber, Andreas; Fuhr, Norbert (Ed.): Advances in Information Retrieval, pp. 817–821, Springer International Publishing, Cham, 2015, ISBN: 978-3-319-16354-3.
@inproceedings{10.1007/978-3-319-16354-3_93,
title = {A Formal Approach to Effectiveness Metrics for Information Access: Retrieval, Filtering, and Clustering},
author = {Enrique Amigó and Julio Gonzalo and Stefano Mizzaro},
editor = {Allan Hanbury and Gabriella Kazai and Andreas Rauber and Norbert Fuhr},
isbn = {978-3-319-16354-3},
year = {2015},
date = {2015-01-01},
booktitle = {Advances in Information Retrieval},
pages = {817–821},
publisher = {Springer International Publishing},
address = {Cham},
abstract = {In this tutorial we present a formal account of evaluation metrics for three of the most salient information related tasks: Retrieval, Clustering, and Filtering. We focus on the most popular metrics and, by exploiting measurement theory, we show some constraints for suitable metrics in each of the three tasks. We also systematically compare metrics according to how they satisfy such constraints, we provide criteria to select the most adequate metric for each specific information access task, and we discuss how to combine and weight metrics.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2014
Amigó, Enrique; Gonzalo, Julio; Mizzaro, Stefano
A general account of effectiveness metrics for information tasks: retrieval, filtering, and clustering Proceedings Article
In: Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 1289, Association for Computing Machinery, Gold Coast, Queensland, Australia, 2014, ISBN: 9781450322577.
@inproceedings{10.1145/2600428.2602296,
title = {A general account of effectiveness metrics for information tasks: retrieval, filtering, and clustering},
author = {Enrique Amigó and Julio Gonzalo and Stefano Mizzaro},
url = {https://doi.org/10.1145/2600428.2602296},
doi = {10.1145/2600428.2602296},
isbn = {9781450322577},
year = {2014},
date = {2014-01-01},
booktitle = {Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval},
pages = {1289},
publisher = {Association for Computing Machinery},
address = {Gold Coast, Queensland, Australia},
series = {SIGIR '14},
abstract = {In this tutorial we will present, review, and compare the most popular evaluation metrics for some of the most salient information related tasks, covering: (i) Information Retrieval, (ii) Clustering, and (iii) Filtering. The tutorial will make a special emphasis on the specification of constraints for suitable metrics in each of the three tasks, and on the systematic comparison of metrics according to such constraints. The last part of the tutorial will investigate the challenge of combining and weighting metrics.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2011
Mizzaro, Stefano; Vassena, Luca
A social approach to context-aware retrieval Journal Article
In: World Wide Web, vol. 14, no 4, pp. 377-405, 2011, ISSN: 1573-1413.
@article{Mizzaro2011,
title = {A social approach to context-aware retrieval},
author = {Stefano Mizzaro and Luca Vassena},
url = {https://doi.org/10.1007/s11280-011-0116-6},
doi = {10.1007/s11280-011-0116-6},
issn = {1573-1413},
year = {2011},
date = {2011-07-01},
journal = {World Wide Web},
volume = {14},
number = {4},
pages = {377-405},
abstract = {We present a general purpose solution to Web content and services perusal by means of mobile devices, named Social Context-Aware Browser. This is a novel approach for information access based on users' context, that exploits social and collaborative models to overtake the limits of the existing solutions. Instead of relying on a pool of experts and on a rigid categorization, as it is usually done in the context-aware field, our solution allows the crowd of users to model, control, and manage the contextual knowledge through collaboration and participation. To have a dynamic and user-tailored context representation, and to enhance the process of retrieval based on users' actual situation, the community of users is encouraged to define the contexts of interest, to share, use, and discuss them, and to associate context to content and resources (Web pages, services, applications, etc.). This paper provides an overall presentation of our solution, describing the idea, the implementation, and the evaluation through a benchmark based methodology.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2010
Coppola, P; Mea, Della V; Gaspero, Di L; Menegon, D; Mischis, D; Mizzaro, S; Scagnetto, I; Vassena, L
The Context-Aware Browser Journal Article
In: IEEE Intelligent Systems, vol. 25, no 1, pp. 38-47, 2010.
@article{5432259,
title = {The Context-Aware Browser},
author = {P Coppola and Della V Mea and Di L Gaspero and D Menegon and D Mischis and S Mizzaro and I Scagnetto and L Vassena},
doi = {10.1109/MIS.2010.26},
year = {2010},
date = {2010-01-01},
journal = {IEEE Intelligent Systems},
volume = {25},
number = {1},
pages = {38-47},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Coppola, Paolo; Mea, Vincenzo Della; Gaspero, Luca Di; Lomuscio, Raffaella; Mischis, Danny; Mizzaro, Stefano; Nazzi, Elena; Scagnetto, Ivan; Vassena, Luca
AI Techniques in a Context-Aware Ubiquitous Environment Book Chapter
In: Hassanien, Aboul-Ella; Abawajy, Jemal H; Abraham, Ajith; Hagras, Hani (Ed.): Pervasive Computing: Innovations in Intelligent Multimedia and Applications, pp. 157–180, Springer London, London, 2010, ISBN: 978-1-84882-599-4.
@inbook{Coppola2010,
title = {AI Techniques in a Context-Aware Ubiquitous Environment},
author = {Paolo Coppola and Vincenzo Della Mea and Luca Di Gaspero and Raffaella Lomuscio and Danny Mischis and Stefano Mizzaro and Elena Nazzi and Ivan Scagnetto and Luca Vassena},
editor = {Aboul-Ella Hassanien and Jemal H Abawajy and Ajith Abraham and Hani Hagras},
url = {https://doi.org/10.1007/978-1-84882-599-4_8},
doi = {10.1007/978-1-84882-599-4_8},
isbn = {978-1-84882-599-4},
year = {2010},
date = {2010-01-01},
booktitle = {Pervasive Computing: Innovations in Intelligent Multimedia and Applications},
pages = {157--180},
publisher = {Springer London},
address = {London},
abstract = {Nowadays, the mobile computing paradigm and the widespread diffusion of mobile devices are quickly changing and replacing many common assumptions about software architectures and interaction/communication models. The environment, in particular, or more generally, the so-called user context is claiming a central role in everyday's use of cellular phones, PDAs, etc. This is due to the huge amount of data ``suggested'' by the surrounding environment that can be helpful in many common tasks. For instance, the current context can help a search engine to refine the set of results in a useful way, providing the user with a more suitable and exploitable information. Moreover, we can take full advantage of this new data source by ``pushing'' active contents towards mobile devices, empowering the latter with new features (e.g., applications) that can allow the user to fruitfully interact with the current context. Following this vision, mobile devices become dynamic self-adapting tools, according to the user needs and the possibilities offered by the environment. The present work proposes MoBe: an approach for providing a basic infrastructure for pervasive context-aware applications on mobile devices, in which AI techniques (namely a principled combination of rule-based systems, Bayesian networks and ontologies) are applied to context inference. The aim is to devise a general inferential framework to make easier the development of context-aware applications by integrating the information coming from physical and logical sensors (e.g., position, agenda) and reasoning about this information in order to infer new and more abstract contexts.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}