88 results on '"Cappellato, L"'
Search Results
2. CLEF 2017 NewsREEL Overview: Contextual Bandit News Recommendation
- Author
-
Liang, Y., Loni, B., Larson, M., Cappellato, L., and Cappellato, L.
- Subjects
CEUR Workshop Proceedings ,Data Science ,ComputingMethodologies_DOCUMENTANDTEXTPROCESSING ,Language & Communication - Abstract
Contains fulltext : 176343.pdf (Publisher’s version ) (Closed access)
- Published
- 2017
3. Cross-Domain Authorship Attribution with Federales
- Author
-
Halteren, H. van, Losada, D.E., Müller, H., Cappellato, L., Ferro, N., Losada, D.E., Müller, H., Cappellato, L., and Ferro, N.
- Subjects
Language & Speech Technology ,Language & Communication - Abstract
Item does not contain fulltext CLEF 2019 - Conference and Labs of the Evaluation Forum, 09 september 2019
- Published
- 2019
4. Bot and Gender Recognition on Tweets using Feature Count Deviations
- Author
-
Halteren, H. van, Losada, D.E., Müller, H., Cappellato, L., Ferro, N., Losada, D.E., Müller, H., Cappellato, L., and Ferro, N.
- Subjects
Language & Speech Technology ,Language & Communication - Abstract
Item does not contain fulltext CLEF 2019 - Conference and Labs of the Evaluation Forum, 09 september 2019
- Published
- 2019
5. Argument Retrieval from Web
- Author
-
Shahshahani, M.S., Kamps, J., Arampatzis, A., Kanoulas, E., Tsikrika, T., Vrochidis, S., Joho, H., Lioma, C., Eickhoff, C., Névéol, A., Cappellato, L., Ferro, N., Language and Computation (ILLC, FNWI/FGw), and ILLC (FGw)
- Subjects
Argumentative ,Search engine ,Information retrieval ,PageRank ,Argument ,Computer science ,law ,Process (engineering) ,Relevance (information retrieval) ,Classifier (UML) ,law.invention ,Task (project management) - Abstract
We are well beyond the days of expecting search engines to help us find documents containing the answer to a question or information about a query. We expect a search engine to help us in the decision-making process. Argument retrieval task in Touché Track at CLEF2020 has been defined to address this problem. The user is looking for information about several alternatives to make a choice between them. The search engine should retrieve opinionated documents containing comparisons between the alternatives rather than documents about one option or documents including personal opinions or no suggestion at all. In this paper, we discuss argument retrieval from web documents. In order to retrieve argumentative documents from the web, we use three features (PageRank scores, domains, argumentative classifier) and try to strike a balance between them. We evaluate the method based on three dimensions: relevance, argumentativeness, and trustworthiness. Since the labeled data and final results for Toucheé Track have not been out yet, the evaluation has been done by manually labeling documents for 5 queries.
- Published
- 2020
6. University of Amsterdam at CLEF 2020: Notebook for the Touché Lab on Argument Retrieval at CLEF 2020
- Author
-
Shahshahani, M.S., Kamps, J., Cappellato, L., Eickhoff, C., Ferro, N., Névéol, A., ILLC (FGw), and Language and Computation (ILLC, FNWI/FGw)
- Subjects
InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL - Abstract
This paper documents the University of Amsterdam’s participation in CLEF 2020 Touché Track. This is the first year this track has been introduced at CLEF, and we were attracted to participate in it due to its potentialities for Parliamentary debates we are currently working on. This track consists of two tasks: Conversational Argument Retrieval and Comparative Argument Retrieval. We submitted a run to both tasks. For the first task, we used a combination of the traditional BM25 model and learning to rank models. BM25 model helps to retrieve relevant arguments, and learning to rank model helps to re-rank the list and put stronger arguments on top of the list. For the second task, Comparative Argument Retrieval, we proposed a pipeline to re-rank documents retrieved from Clueweb using three features: PageRank scores, web domains, and argumentativeness. Preliminary results on 5 queries have shown that this heuristic pipeline may help to achieve a balance among three important dimensions: relevance, trustworthiness, and argumentativeness.
- Published
- 2020
7. Named Entity Recognition and Linking on Historical Newspapers: UvA.ILPS & REL at CLEF HIPE 2020
- Author
-
Cappellato, L., Provatorova, V., Vakulenko, S., Kanoulas, E., Dercksen, K., Hulst, J.M. van, Cappellato, L., Provatorova, V., Vakulenko, S., Kanoulas, E., Dercksen, K., and Hulst, J.M. van
- Abstract
CLEF 2020, Contains fulltext : 233830.pdf (Publisher’s version ) (Open Access)
- Published
- 2020
8. Overview of CLEF HIPE 2020: Named Entity Recognition and Linking on Historical Newspapers
- Author
-
Arampatzis, Avi, Kanoulas, Evangelos, Tsikrika, Theodora, Vrochidis, Stefanos, Joho, Hideo, Lioma, Christina, Eickhoff, Carsten, Névéol, Aurélie, Cappellato, Linda, Ferro, Nicola, Arampatzis, A ( Avi ), Kanoulas, E ( Evangelos ), Tsikrika, T ( Theodora ), Vrochidis, S ( Stefanos ), Joho, H ( Hideo ), Lioma, C ( Christina ), Eickhoff, C ( Carsten ), Névéol, A ( Aurélie ), Cappellato, L ( Linda ), Ferro, N ( Nicola ), Ehrmann, Maud; https://orcid.org/0000-0001-9900-2193, Romanello, Matteo; https://orcid.org/0000-0002-1890-2577, Flückiger, Alex, Clematide, Simon; https://orcid.org/0000-0003-1365-0662, Arampatzis, Avi, Kanoulas, Evangelos, Tsikrika, Theodora, Vrochidis, Stefanos, Joho, Hideo, Lioma, Christina, Eickhoff, Carsten, Névéol, Aurélie, Cappellato, Linda, Ferro, Nicola, Arampatzis, A ( Avi ), Kanoulas, E ( Evangelos ), Tsikrika, T ( Theodora ), Vrochidis, S ( Stefanos ), Joho, H ( Hideo ), Lioma, C ( Christina ), Eickhoff, C ( Carsten ), Névéol, A ( Aurélie ), Cappellato, L ( Linda ), Ferro, N ( Nicola ), Ehrmann, Maud; https://orcid.org/0000-0001-9900-2193, Romanello, Matteo; https://orcid.org/0000-0002-1890-2577, Flückiger, Alex, and Clematide, Simon; https://orcid.org/0000-0003-1365-0662
- Abstract
This paper presents an overview of the first edition of HIPE (Identifying Historical People, Places and other Entities), a pioneering shared task dedicated to the evaluation of named entity processing on historical newspapers in French, German and English. Since its introduction some twenty years ago, named entity (NE) processing has become an essential component of virtually any text mining application and has undergone major changes. Recently, two main trends characterise its developments: the adoption of deep learning architectures and the consideration of textual material originating from historical and cultural heritage collections. While the former opens up new opportunities, the latter introduces new challenges with heterogeneous, historical and noisy inputs. In this context, the objective of HIPE, run as part of the CLEF 2020 conference, is threefold: strengthening the robustness of existing approaches on non-standard inputs, enabling performance comparison of NE processing on historical texts, and, in the long run, fostering efficient semantic indexing of historical documents. Tasks, corpora, and results of 13 participating teams are presented.
- Published
- 2020
9. Overview of the CLEF eHealth Evaluation Lab 2019
- Author
-
Kelly, L., Suominen, H., Goeuriot, L., Neves, M., Kanoulas, E., Li, D., Azzopardi, L., Spijker, R., Zuccon, G., Scells, H., Palotti, J., Crestani, F., Braschler, M., Savoy, J., Rauber, A., Müller, H., Losada, D.E., Heinatz Bürki, G., Cappellato, L., Ferro, N., Computer Science Department [Maynooth], National University of Ireland Maynooth (Maynooth University), Data61 [Canberra] (CSIRO), Australian National University (ANU)-Commonwealth Scientific and Industrial Research Organisation [Canberra] (CSIRO), Australian National University (ANU), University of Turku, Laboratoire d'Informatique de Grenoble (LIG), Centre National de la Recherche Scientifique (CNRS)-Université Grenoble Alpes (UGA)-Institut polytechnique de Grenoble - Grenoble Institute of Technology (Grenoble INP ), Université Grenoble Alpes (UGA), German Centre for the Protection of Laboratory Animals (Bf3R), Informatics Institute [Amsterdam], University of Amsterdam [Amsterdam] (UvA), Department of Computer and Information Sciences [Univ Strathclyde], University of Strathclyde [Glasgow], Cochrane Netherlands and UMC Utrecht, University of Queensland [Brisbane], Qatar Computing Research Institute [Doha, Qatar] (QCRI), Information and Language Processing Syst (IVI, FNWI), and IVI (FNWI)
- Subjects
0303 health sciences ,020205 medical informatics ,Computer science ,business.industry ,02 engineering and technology ,computer.software_genre ,Health informatics ,Clef ,3. Good health ,Task (project management) ,World Wide Web ,03 medical and health sciences ,Information extraction ,Entity linking ,Resource (project management) ,Systematic review ,[INFO.INFO-IR]Computer Science [cs]/Information Retrieval [cs.IR] ,0202 electrical engineering, electronic engineering, information engineering ,eHealth ,business ,computer ,ComputingMilieux_MISCELLANEOUS ,030304 developmental biology - Abstract
In this paper, we provide an overview of the seventh annual edition of the CLEF eHealth evaluation lab. CLEF eHealth 2019 continues our evaluation resource building efforts around the easing and support of patients, their next-of-kins, clinical staff, and health scientists in understanding, accessing, and authoring electronic health information in a multilingual setting. This year’s lab advertised three tasks: Task 1 on indexing non-technical summaries of German animal experiments with International Classification of Diseases, Version 10 codes; Task 2 on technology assisted reviews in empirical medicine building on 2017 and 2018 tasks in English; and Task 3 on consumer health search in mono- and multilingual settings that builds on the 2013–18 Information Retrieval tasks. In total nine teams took part in these tasks (six in Task 1 and three in Task 2). Herein, we describe the resources created for these tasks and evaluation methodology adopted. We also provide a brief summary of participants of this year’s challenges and results obtained. As in previous years, the organizers have made data and tools associated with the lab tasks available for future research and development.
- Published
- 2019
- Full Text
- View/download PDF
10. Automatic thresholding by sampling documents and estimating recall: ILPs@UVA at Tar task 2.2
- Author
-
Li, D., Kanoulas, E., Cappellato, L., Ferro, N., Losada, D.E., Müller, H., Information and Language Processing Syst (IVI, FNWI), and Faculty of Science
- Abstract
In this paper, we describe the participation of the Information and Language Processing System (ILPS) group at CLEF eHealth 2019 Task 2.2: Technologically Assisted Reviews in Empirical Medicine. This task is targeted to produce an efficient ordering of the documents and to identify a subset of the documents which contains as many of the relevant abstracts for the least effort. Participants are provided with systematic review topics with each including a review title, a boolean query constructed by Cochrane experts, and a set of PubMed Document Identifiers (PID's) returned by running the boolean query in MEDLINE. We handle the problem under the Continuous Active Learning framework by jointly training a ranking model to rank documents, and conducting a “greedy” sampling to estimate the real number of relevant documents in the collection. We finally submitted four runs.
- Published
- 2019
11. Cross-Domain Authorship Attribution with Federales
- Author
-
Losada, D.E., Müller, H., Cappellato, L., Ferro, N., Halteren, H. van, Losada, D.E., Müller, H., Cappellato, L., Ferro, N., and Halteren, H. van
- Abstract
CLEF 2019 - Conference and Labs of the Evaluation Forum, 9 september 2019, Item does not contain fulltext
- Published
- 2019
12. Bot and Gender Recognition on Tweets using Feature Count Deviations
- Author
-
Losada, D.E., Müller, H., Cappellato, L., Ferro, N., Halteren, H. van, Losada, D.E., Müller, H., Cappellato, L., Ferro, N., and Halteren, H. van
- Abstract
CLEF 2019 - Conference and Labs of the Evaluation Forum, 9 september 2019, Item does not contain fulltext
- Published
- 2019
13. Celebrity profiling on twitter using sociolinguistic features notebook for PAN at CLEF 2019
- Author
-
Cappellato L., Ferro N., Losada D.E., Muller H., Moreno-Sandoval L.G., Puertas E., Plaza-Del-Arco F.M., Pomares-Quimbaya A., Alvarado-Valencia J.A., Alfonso Ureña-López L., Cappellato L., Ferro N., Losada D.E., Muller H., Moreno-Sandoval L.G., Puertas E., Plaza-Del-Arco F.M., Pomares-Quimbaya A., Alvarado-Valencia J.A., and Alfonso Ureña-López L.
- Abstract
Social networks have been a revolutionary scenario for celebrities because they allow them to reach a wider audience with much higher frequency than using traditional means. These platforms enable them to improve or sometimes deteriorate, their careers through the construction of closer relationships with their fans and the acquisition of new ones. Indeed, networks have promoted the emergence of a new type of celebrities that exists only in the digital world. Being able to characterize the celebrities that are more active on social networks, such as Twitter, gives an enormous opportunity to identify what is their real level of fame, what is their relevance for an age group, or a specific gender or occupation. These facts may enrich decision making, especially in advertising and marketing. To achieve this aim, this paper presents a novel strategy for the characterization of celebrities profile on Twitter based on the generation of socio-linguistic features from their posts that serve as input to a set of classifiers. Specifically, we produced four classifiers that describe the level of fame, the gender, the birth date, and the possible occupation of a celebrity. We obtained the training and test data sets as part of our participation at PAN 2019 at CLEF. Results of each classifier are reported including the analysis of which features are more relevant, which classification techniques were more useful and which were the final precision and recall results. © 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
- Published
- 2019
14. Bots and gender profiling on twitter using sociolinguistic features notebook for PAN at CLEF 2019
- Author
-
Cappellato L., Ferro N., Losada D.E., Muller H., Puertas E., Moreno-Sandoval L.G., Plaza-Del-Arco F.M., Alvarado‑Valencia, Jorge Andres, Pomares-Quimbaya A., Alfonso Ureña-López L., Cappellato L., Ferro N., Losada D.E., Muller H., Puertas E., Moreno-Sandoval L.G., Plaza-Del-Arco F.M., Alvarado‑Valencia, Jorge Andres, Pomares-Quimbaya A., and Alfonso Ureña-López L.
- Abstract
Unfortunately, in social networks, software bots or just bots are becoming more and more common because malicious people have seen their usefulness to spread false messages, spread rumors and even manipulate public opinion. Even though the text generated by users in social networks is a rich source of information that can be used to identify different aspects of its authors, not being able to recognize which users are truly humans and which are not, is a big drawback. In this work, we describe the properties of our multilingual classification model submitted for PAN2019 that is able to recognize bots from humans, and females from males. This solution extracted 18 features from the user's posts and applying a machine learning algorithm obtained good performance results. © 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
- Published
- 2019
15. Overview of the CLEF 2019 Personalised Information Retrieval Lab (PIR-CLEF 2019)
- Author
-
Crestani, F, Braschler, M, Savoy, J, Rauber, A, Muller, H, Losada, DE, Burki, GH, Cappellato, L, Ferro, N, Pasi, G, Jones, G, Goeuriot, L, Kelly, L, Marrara, S, Sanvitto, C, Jones, GJF, Crestani, F, Braschler, M, Savoy, J, Rauber, A, Muller, H, Losada, DE, Burki, GH, Cappellato, L, Ferro, N, Pasi, G, Jones, G, Goeuriot, L, Kelly, L, Marrara, S, Sanvitto, C, and Jones, GJF
- Abstract
The Personalised Information Retrieval Lab (PIR-CLEF 2019) lab is an initiative aimed at both providing and critically analysing the evaluation of Personalization in Information Retrieval (PIR) applications. PIR-CLEF 2019 is the second edition of the Lab after the successful Pilot lab organised at CLEF 2017 and the first edition of the Lab at CLEF 2018. PIR-CLEF 2019 provided registered participants with two tracks: the Web Search Task and the Medical Search Task. The Web Search Task continues the activities introduced in the previous editions of the PIR-CLEF Lab, while the Medical Search Track focuses on personalisation within an ad hoc search task introduced in previous editions of the CLEF eHealth Lab.
- Published
- 2019
16. CLEF 2017 NewsREEL Overview: A Stream-based Recommender Task for Evaluation and Education
- Author
-
Lommatzsch, A., Kille, B., Hopfgartner, F., Larson, M., Brodt, T., Seiler, J., Özgöbek, Ö., Jones, G.J.F., Lawless, S., Gonzalo, J., Kelly, L., Goeuriot, L., Mandl, T., Cappellato, L., Ferro, N., Jones, Gareth J.F., Lawless, Seamus, Gonzalo, Julio, Kelly, Liadh, Goeuriot, Lorraine, Mandl, Thomas, Cappellato, Linda, Ferro, Nicola, Jones, G.J.F., Lawless, S., Gonzalo, J., Kelly, L., Goeuriot, L., Mandl, T., Cappellato, L., and Ferro, N.
- Subjects
Data Science ,Language & Communication - Abstract
News recommender systems provide users with access to news stories that they find interesting and relevant. As other online, stream-based recommender systems, they face particular challenges, including limited information on users’ preferences and also rapidly fluctuating item collections. In addition, technical aspects, such as response time and scalability, must be considered. Both algorithmic and technical considerations shape working requirements for real-world recommender systems in businesses. NewsREEL represents a unique opportunity to evaluate recommendation algorithms and for students to experience realistic conditions and to enlarge their skill sets. The NewsREEL Challenge requires participants to conduct data-driven experiments in NewsREEL Replay as well as deploy their best models into NewsREEL Live’s ‘living lab’. This paper presents NewsREEL 2017 and also provides insights into the effectiveness of NewsREEL to support the goals of instructors teaching recommender systems to students. We discuss the experiences of NewsREEL participants as well as those of instructors teaching recommender systems to students, and in this way, we showcase NewsREEL’s ability to support the education of future data scientists.
- Published
- 2017
17. Simply the Best: Minimalist System Trumps Complex Models in Author Profiling
- Author
-
Basile, Angelo, Dwyer, Gareth, Medvedeva, Maria, Rawee, Josine, Haagsma, Hessel, Nissim, Malvina, Bellot, P., Trabelsi, C., Mothe, J., Murtagh, F., Nie, J. Y., Soulier, L., SanJuan, E., Cappellato, L., and Ferro, N.
- Abstract
A simple linear SVM with word and character n-gram features and minimal parameter tuning can identify the gender and the language variety (for English, Spanish, Arabic and Portuguese) of Twitter users with very high accuracy. All our attempts at improving performance by including more data, smarter features, and employing more complex architectures plainly fail. In addition, we experiment with joint and multitask modelling, but find that they are clearly outperformed by single task models. Eventually, our simplest model was submitted to the PAN 2017 shared task on author profiling, obtaining an average accuracy of 0.86 on the test set, with performance on sub-tasks ranging from 0.68 to 0.98. These were the best results achieved at the competition overall. To allow lay people to easily use and see the value of machine learning for author profiling, we also built a web application on top our models.
- Published
- 2018
18. Overview of the CLEF dynamic search evaluation lab 2018
- Author
-
Kanoulas, E., Azzopardi, L., Yang, G.H., Bellot, P., Trabelsi, C., Mothe, J., Murtagh, F., Nie, J.Y., Soulier, L., SanJuan, E., Cappellato, L., Ferro, N., Information and Language Processing Syst (IVI, FNWI), and Operations Management (ABS, FEB)
- Subjects
Information retrieval ,Computer science ,Contextual advertising ,02 engineering and technology ,Clef ,Session (web analytics) ,Personalization ,Ranking (information retrieval) ,Search algorithm ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Dialog box ,Metasearch engine - Abstract
In this paper we provide an overview of the CLEF 2018 Dynamic Search Lab. The lab ran for the first time in 2017 as a workshop. The outcomes of the workshop were used to define the tasks of this year’s evaluation lab. The lab strives to answer one key question: how can we evaluate, and consequently build, dynamic search algorithms? Unlike static search algorithms, which consider user request’s independently, and consequently do not adapt their ranking with respect to the user’s sequence of interactions and the user’s end goal, dynamic search algorithms try to infer the user’s intentions based on their interactions and adapt their ranking accordingly. Session personalization, contextual search, conversational search, dialog systems are some examples of dynamic search. Herein, we describe the overall objectives of the CLEF 2018 Dynamic Search Lab, the resources created, and the evaluation methodology designed.
- Published
- 2018
- Full Text
- View/download PDF
19. CLEF 2018 technologically assisted reviews in empirical medicine overview
- Author
-
Kanoulas, E., Li, D., Azzopardi, L., Spijker, R., Cappellato, L., Ferro, N., Nie, J.-Y., Soulier, L., Operations Management (ABS, FEB), Information and Language Processing Syst (IVI, FNWI), and IvI Research (FNWI)
- Abstract
Conducting a systematic review is a widely used method to obtain an overview over the current scientific consensus on a topic of interest, by bringing together multiple studies in a reliable, transparent way. The large and growing number of published studies, and their increasing rate of publication, makes the task of identifying all relevant studies in an unbiased way both complex and time consuming to the extent that jeopardizes the validity of their findings and the ability to inform policy and practice in a timely manner. The CLEF 2018 e-Health Technology Assisted Reviews in Empirical Medicine task aims at evaluating search algorithms that seek to identify all studies relevant for conducting a systematic review in empirical medicine. The task had a focus on Diagnostic Test Accuracy (DTA) reviews, and consisted of two subtasks: 1) given a number of relevance criteria as described in a systematic review protocol, search a large medical database of article abstracts (PubMed) to find the studies to be included in the review, and 2) given the article abstracts retrieved by a carefully designed Boolean Query, prioritize them to reduce the effort required by experts to screen the abstracts for inclusion in the review. Seven teams participated in the task, with a total of 12 runs submitted for subtask 1 and 19 runs for subtask 2. This paper reports both the methodology used to construct the benchmark collection, and the results of the evaluation.
- Published
- 2018
20. Overview of the CLEF eHealth Evaluation Lab 2018
- Author
-
Suominen, H., Kelly, L., Goeuriot, L., Névéol, A., Ramadier, L., Robert, A., Kanoulas, E., Spijker, R., Azzopardi, L., Li, D., Jimmy, Palotti, J., Zuccon, G., Bellot, P., Trabelsi, C., Mothe, J., Murtagh, F., Nie, J.Y., Soulier, L., SanJuan, E., Cappellato, L., Ferro, N., Information and Language Processing Syst (IVI, FNWI), Operations Management (ABS, FEB), IvI Research (FNWI), SanJuan, Eric, Murtagh, Fionn, Nie, Jian Yun, Soulier, Laure, Cappellato, Linda, Bellot, Patrice, Mothe, Josiane, Trabelsi, Chiraz, and Ferro, Nicola
- Subjects
QA75 ,Computer science ,business.industry ,02 engineering and technology ,computer.software_genre ,Health informatics ,Clef ,Task (project management) ,World Wide Web ,Information extraction ,Entity linking ,Systematic review ,Resource (project management) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,eHealth ,020201 artificial intelligence & image processing ,business ,computer - Abstract
In this paper, we provide an overview of the sixth annual edition of the CLEF eHealth evaluation lab. CLEF eHealth 2018 continues our evaluation resource building efforts around the easing and support of patients, their next-of-kins, clinical staff, and health scientists in understanding, accessing, and authoring eHealth information in a multilingual setting. This year’s lab offered three tasks: Task 1 on multilingual information extraction to extend from last year’s task on French and English corpora to French, Hungarian, and Italian; Task 2 on technologically assisted reviews in empirical medicine building on last year’s pilot task in English; and Task 3 on Consumer Health Search (CHS) in mono- and multilingual settings that builds on the 2013–17 Information Retrieval tasks. In total 28 teams took part in these tasks (14 in Task 1, 7 in Task 2 and 7 in Task 3). Herein, we describe the resources created for these tasks, outline our evaluation methodology adopted and provide a brief summary of participants of this year’s challenges and results obtained. As in previous years, the organizers have made data and tools associated with the lab tasks available for future research and development.
- Published
- 2018
- Full Text
- View/download PDF
21. Overview of the Living Labs for Information Retrieval Evaluation (LL4IR) CLEF Lab 2015
- Author
-
Schuth, A., Balog, K., Kelly, L., Mothe, J., Savoy, J., Kamps, J., Pinel-Sauvagnat, K., Jones, G.J.F., SanJuan, E., Cappellato, L., Ferro, N., and Information and Language Processing Syst (IVI, FNWI)
- Subjects
World Wide Web ,Information retrieval ,Computer science ,Benchmarking ,Clef ,Task (project management) ,Production system - Abstract
In this paper we report on the first Living Labs for Information Retrieval Evaluation (LL4IR) CLEF Lab. Our main goal with the lab is to provide a benchmarking platform for researchers to evaluate their ranking systems in a live setting with real users in their natural task environments. For this first edition of the challenge we focused on two specific use-cases: product search and web search. Ranking systems submitted by participants were experimentally compared using interleaved comparisons to the production system from the corresponding use-case. In this paper we describe how these experiments were performed, what the resulting outcomes are, and conclude with some lessons learned.
- Published
- 2015
- Full Text
- View/download PDF
22. Overview of the CLEF 2018 Personalised Information Retrieval Lab (PIR-CLEF 2018)
- Author
-
Cappellato, L, Ferro, N, Nie, JY, Soulier, L, Pasi, G, Jones, G, Curtis, K, Marrara, S, Sanvitto, C, Ganguly, D, Sen, P, Pasi G., Jones G. J. F., Curtis K., Marrara S., Sanvitto C., Ganguly D., Sen P., Cappellato, L, Ferro, N, Nie, JY, Soulier, L, Pasi, G, Jones, G, Curtis, K, Marrara, S, Sanvitto, C, Ganguly, D, Sen, P, Pasi G., Jones G. J. F., Curtis K., Marrara S., Sanvitto C., Ganguly D., and Sen P.
- Abstract
At CLEF 2018, the Personalised Information Retrieval Lab (PIR-CLEF 2018) has been conceived to provide an initiative aimed at both providing and critically analysing a new approach to the evaluation of personalization in Information Retrieval (PIR). PIR-CLEF 2018 is the first edition of this Lab after the successful Pilot lab organised at CLEF 2017. PIR CLEF 2018 has provided registered participants with the data sets originally developed for the PIR-CLEF 2017 Pilot task; the data collected are related to real search sessions over a subset of the ClueWebl2 collection, undertaken by 10 users by using a novel methodology. The data were gathered during the search sessions undertaken by 10 volunteer searchers. Activities during these search sessions included relevance assessment of a retrieved documents by the searchers. 16 groups registered to participate at PIR-CLEF 2018 and were provided with the data set to allow them to work on PIR related tasks and to provide feedback about our proposed PIR evaluation methodology with the aim to create an effective evaluation task.
- Published
- 2018
23. Evaluation of personalised information retrieval at CLEF 2018 (PIR-CLEF)
- Author
-
Bellot, P, Trabelsi, C, Mothe, J, Murtagh, F, Nie, JY, Soulier, L, SanJuan, E, Cappellato, L, Ferro, N, Pasi, G, Jones, G, Curtis, K, Marrara, S, Sanvitto, C, Ganguly, D, Sen, P, Pasi, Gabriella, Jones, Gareth J. F., Curtis, Keith, Marrara, Stefania, Sanvitto, Camilla, Ganguly, Debasis, Sen, Procheta, Bellot, P, Trabelsi, C, Mothe, J, Murtagh, F, Nie, JY, Soulier, L, SanJuan, E, Cappellato, L, Ferro, N, Pasi, G, Jones, G, Curtis, K, Marrara, S, Sanvitto, C, Ganguly, D, Sen, P, Pasi, Gabriella, Jones, Gareth J. F., Curtis, Keith, Marrara, Stefania, Sanvitto, Camilla, Ganguly, Debasis, and Sen, Procheta
- Abstract
The series of Personalised Information Retrieval (PIR-CLEF) Labs at CLEF is intended as a forum for the exploration of methodologies for the repeatable evaluation of personalised information retrieval (PIR). The PIR-CLEF 2018 Lab is the first full edition of this series after the successful pilot edition at CLEF 2017, and provides a Lab task dedicated to personalised search, while the workshop at the conference will form the basis of further discussion of strategies for the evaluation of PIR and suggestions for improving the activities of the PIR-CLEF Lab. The PIR-CLEF 2018 Task is the first PIR evaluation benchmark based on the Cranfield paradigm, with the potential benefits of producing evaluation results that are easily reproducible. The task is based on search sessions over a subset of the ClueWeb12 collection, undertaken by volunteer searchers using a methodology developed in the CLEF 2017 pilot edition of PIR-CLEF. The PIR-CLEF test collection provides a detailed set of data gathered during the activities undertaken by each subject during the search sessions, including their search queries and details of relevant documents as marked by the searchers. The PIR-CLEF 2018 workshop is intended to review the design and construction of the collection, and to consider the topic of reproducible evaluation of PIR more generally with the aim of improving future editions of the evaluation benchmark.
- Published
- 2018
24. CLEF 2017 dynamic search evaluation lab overview
- Author
-
Kanoulas, E., Azzopardi, L., Jones, G.J.F., Lawless, S., Gonzalo, J., Kelly, L., Goeuriot, L., Mandl, T., Cappellato, L., Ferro, N., Faculty of Science, IvI Research (FNWI), IVI (FNWI), Jones, Gareth J.F., Lawless, Séamus, Gonzalo, Julio, Kelly, Liadh, Goeuriot, Lorraine, Mandl, Thomas, Cappellato, Linda, and Ferro, Nicola
- Subjects
QA75 ,Information retrieval ,Computer science ,Context (language use) ,Contextual advertising ,02 engineering and technology ,Session (web analytics) ,Clef ,Ranking (information retrieval) ,Task (project management) ,Search algorithm ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Key (cryptography) ,020201 artificial intelligence & image processing - Abstract
In this paper we provide an overview of the first edition of the CLEF Dynamic Search Lab. The CLEF Dynamic Search lab ran in the form of a workshop with the goal of approaching one key question: how can we evaluate dynamic search algorithms? Unlike static search algorithms, which essentially consider user request’s independently, and which do not adapt the ranking w.r.t the user’s sequence of interactions, dynamic search algorithms try to infer from the user’s intentions from their interactions and then adapt the ranking accordingly. Personalized session search, contextual search, and dialog systems often adopt such algorithms. This lab provides an opportunity for researchers to discuss the challenges faced when trying to measure and evaluate the performance of dynamic search algorithms, given the context of available corpora, simulations methods, and current evaluation metrics. To seed the discussion, a pilot task was run with the goal of producing search agents that could simulate the process of a user, interacting with a search system over the course of a search session. Herein, we describe the overall objectives of the CLEF 2017 Dynamic Search Lab, the resources created for the pilot task and the evaluation methodology adopted.
- Published
- 2017
25. CLEF 2017 Dynamic Search Lab Overview And Evaluation
- Author
-
Kanoulas, E., Azzopardi, L., Cappellato, L., Ferro, N., Goeuriot, L., Mandl, T., and IvI Research (FNWI)
- Published
- 2017
26. CLEF 2017 Technologically Assisted Reviews in Empirical Medicine Overview
- Author
-
Kanoulas, E., Li, D., Azzopardi, L., Spijker, R., Cappellato, L., Ferro, N., Goeuriot, L., Mandl, T., IvI Research (FNWI), and Information and Language Processing Syst (IVI, FNWI)
- Published
- 2017
27. CLEF 2017 eHealth evaluation lab overview
- Author
-
Goeuriot, L., Kelly, L., Suominen, H., Névéol, A., Robert, A., Kanoulas, E., Spijker, R., Palotti, J., Zuccon, G., Jones, G.J.F., Lawless, S., Gonzalo, J., Mandl, T., Cappellato, L., Ferro, N., and IvI Research (FNWI)
- Subjects
Multimedia ,Computer science ,business.industry ,05 social sciences ,computer.software_genre ,Health informatics ,Clef ,Task (project management) ,World Wide Web ,03 medical and health sciences ,Entity linking ,Information extraction ,0302 clinical medicine ,Resource (project management) ,Systematic review ,eHealth ,030212 general & internal medicine ,0509 other social sciences ,050904 information & library sciences ,business ,computer - Abstract
In this paper we provide an overview of the fifth edition of the CLEF eHealth evaluation lab. CLEF eHealth 2017 continues our evaluation resource building efforts around the easing and support of patients, their next-of-kins, clinical staff, and health scientists in understanding, accessing, and authoring eHealth information in a multilingual setting. This year’s lab offered three tasks: Task 1 on multilingual information extraction to extend from last year’s task on French corpora, Task 2 on technologically assisted reviews in empirical medicine as a new pilot task, and Task 3 on patient-centered information retrieval (IR) building on the 2013-16 IR tasks. In total 32 teams took part in these tasks (11 in Task 1, 14 in Task 2, and 7 in Task 3). We also continued the replication track from 2016. Herein, we describe the resources created for these tasks, evaluation methodology adopted and provide a brief summary of participants of this year’s challenges and results obtained. As in previous years, the organizers have made data and tools associated with the lab tasks available for future research and development.
- Published
- 2017
28. CLEF 2017 NewsREEL Overview: A Stream-Based Recommender Task for Evaluation and Education
- Author
-
Jones, G.J.F., Lawless, S., Gonzalo, J., Kelly, L., Goeuriot, L., Mandl, T., Cappellato, L., Ferro, N., Lommatzsch, A., Kille, B., Hopfgartner, F., Larson, M., Brodt, T., Seiler, J., Özgöbek, Ö., Jones, G.J.F., Lawless, S., Gonzalo, J., Kelly, L., Goeuriot, L., Mandl, T., Cappellato, L., Ferro, N., Lommatzsch, A., Kille, B., Hopfgartner, F., Larson, M., Brodt, T., Seiler, J., and Özgöbek, Ö.
- Abstract
Item does not contain fulltext
- Published
- 2017
29. CLEF 2017 NewsREEL Overview: Contextual Bandit News Recommendation
- Author
-
Cappellato, L., Liang, Y., Loni, B., Larson, M., Cappellato, L., Liang, Y., Loni, B., and Larson, M.
- Abstract
Contains fulltext : 176343.pdf (publisher's version ) (Closed access)
- Published
- 2017
30. CLEF 2017 NewsREEL Overview: Offline and Online Evaluation of Stream-based News Recommender Systems
- Author
-
Cappellato, L., Kille, B., Lommatzsch, A., Hopfgartner, F., Larson, M., Brodt, T., Cappellato, L., Kille, B., Lommatzsch, A., Hopfgartner, F., Larson, M., and Brodt, T.
- Abstract
Contains fulltext : 176354.pdf (publisher's version ) (Closed access)
- Published
- 2017
31. QUT ielab at CLEF 2017 e-Health IR task: Knowledge base retrieval for consumer health search
- Author
-
Ferro, N, Mandl, T, Goeuriot, L, Cappellato, L, Jimmy, Jimmy, Zuccon, Guido, Koopman, Bevan, Ferro, N, Mandl, T, Goeuriot, L, Cappellato, L, Jimmy, Jimmy, Zuccon, Guido, and Koopman, Bevan
- Published
- 2017
32. CLEF 2017 Task overview: The IR Task at the eHealth Evaluation Lab: Evaluating retrieval methods for consumer Health search
- Author
-
Ferro, N, Mandl, T, Goeuriot, L, Cappellato, L, Palotti, Joao, Zuccon, Guido, Jimmy, Jimmy, Pecina, Pavel, Lupu, Mihai, Goeuriot, Lorraine, Kelly, Liadh, Hanbury, Allan, Ferro, N, Mandl, T, Goeuriot, L, Cappellato, L, Palotti, Joao, Zuccon, Guido, Jimmy, Jimmy, Pecina, Pavel, Lupu, Mihai, Goeuriot, Lorraine, Kelly, Liadh, and Hanbury, Allan
- Published
- 2017
33. CLEF 2017 eHealth Evaluation Lab overview
- Author
-
Jones, G J F, Kelly, L, Mandl, T, ferro, N, Lawless, S, Gonzalo, J, Goeuriot, L, Cappellato, L, Goeuriot, Lorraine, Kelly, Liadh, Suominen, Hanna, Neveol, Aurelie, Robert, Aude, Kanoulas, Evangelos, Spijker, Rene, Palotti, Joao, Zuccon, Guido, Jones, G J F, Kelly, L, Mandl, T, ferro, N, Lawless, S, Gonzalo, J, Goeuriot, L, Cappellato, L, Goeuriot, Lorraine, Kelly, Liadh, Suominen, Hanna, Neveol, Aurelie, Robert, Aude, Kanoulas, Evangelos, Spijker, Rene, Palotti, Joao, and Zuccon, Guido
- Abstract
In this paper we provide an overview of the fifth edition of the CLEF eHealth evaluation lab. CLEF eHealth 2017 continues our evaluation resource building efforts around the easing and support of patients, their next-of-kins, clinical staff, and health scientists in understanding, accessing, and authoring eHealth information in a multilingual setting. This year’s lab offered three tasks: Task 1 on multilingual information extraction to extend from last year’s task on French corpora, Task 2 on technologically assisted reviews in empirical medicine as a new pilot task, and Task 3 on patient-centered information retrieval (IR) building on the 2013-16 IR tasks. In total 32 teams took part in these tasks (11 in Task 1, 14 in Task 2, and 7 in Task 3). We also continued the replication track from 2016. Herein, we describe the resources created for these tasks, evaluation methodology adopted and provide a brief summary of participants of this year’s challenges and results obtained. As in previous years, the organizers have made data and tools associated with the lab tasks available for future research and development.
- Published
- 2017
34. QUT ielab at CLEF eHealth 2017 Technology Assisted Reviews Track: Initial experiments with learning to rank
- Author
-
Ferro, N, Mandl, T, Goeuriot, L, Cappellato, L, Scells, Harrisen, Zuccon, Guido, Deacon, Anthony, Koopman, Bevan, Ferro, N, Mandl, T, Goeuriot, L, Cappellato, L, Scells, Harrisen, Zuccon, Guido, Deacon, Anthony, and Koopman, Bevan
- Abstract
In this paper we describe our participation to the CLEF eHealth 2017 Technology Assisted Reviews track (TAR). This track aims to evaluate and advance search technologies aimed at supporting the creation of biomedical systematic reviews. In this context, the track explores the task of screening prioritisation: the ranking of studies to be screened for inclusion in a systematic review. Our solution addresses this challenge by developing ranking strategies based on learning to rank techniques and exploiting features derived by the use of the PICO framework. PICO (Population, Intervention, Control or comparison and Outcome) is a technique used in evidence based practice to frame and answer clinical questions and is used extensively in the compilation of systematic reviews. Our experiments show that the use of the PICO-based feature within learning to rank provides improvements over the use of baseline features alone.
- Published
- 2017
35. A language-modelling approach to User-Centred Health Information Retrieval
- Author
-
Verberne, S., Cappellato, L., and Cappellato, L.
- Subjects
CEUR Workshop Proceedings ,Data Science ,ComputingMethodologies_DOCUMENTANDTEXTPROCESSING - Abstract
Contains fulltext : 132752.pdf (Author’s version preprint ) (Open Access)
- Published
- 2014
36. Distributional Semantics for Medical Information Extraction
- Author
-
Quiroz, L., Mennes, L., Dehghani, M., Kanoulas, E., Balog, K., Cappellato, L., Ferro, N., Macdonald, C., and Information and Language Processing Syst (IVI, FNWI)
- Published
- 2016
37. Two-Way Parsimonious Classification Models for Evolving Hierarchies
- Author
-
Dehghani, M., Azarbonyad, H., Kamps, J., Marx, M., Fuhr, N., Quaresma, P., Gonçalves, T., Larsen, B., Balog, K., Macdonald, C., Cappellato, L., Ferro, N., Language and Computation (ILLC, FNWI/FGw), ILLC (FGw), Cultural Heritage and Identity, Faculteit der Geesteswetenschappen, and IvI Research (FNWI)
- Subjects
Hierarchy ,Computer science ,Parliament ,business.industry ,Dynamic data ,media_common.quotation_subject ,Knowledge economy ,02 engineering and technology ,Data structure ,computer.software_genre ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Graph (abstract data type) ,020201 artificial intelligence & image processing ,Artificial intelligence ,Language model ,business ,computer ,Natural language processing ,media_common - Abstract
There is an increasing volume of semantically annotated data available, in particular due to the emerging use of knowledge bases to annotate or classify dynamic data on the web. This is challenging as these knowledge bases have a dynamic hierarchical or graph structure demanding robustness against changes in the data structure over time. In general, this requires us to develop appropriate models for the hierarchical classes that capture all, and only, the essential solid features of the classes which remain valid even as the structure changes. We propose hierarchical significant words language models of textual objects in the intermediate levels of hierarchies as robust models for hierarchical classification by taking the hierarchical relations into consideration. We conduct extensive experiments on richly annotated parliamentary proceedings linking every speech to the respective speaker, their political party, and their role in the parliament. Our main findings are the following. First, we define hierarchical significant words language models as an iterative estimation process across the hierarchy, resulting in tiny models capturing only well grounded text features at each level. Second, we apply the resulting models to party membership and party position classification across time periods, where the structure of the parliament changes, and see the models dramatically better transfer across time periods, relative to the baselines.
- Published
- 2016
- Full Text
- View/download PDF
38. Overview of the CLEF 2016 Social Book Search Lab
- Author
-
Koolen, M., Bogers, T., Gäde, M., Hall, M., Hendrickx, I., Huurdeman, H., Kamps, J., Skov, M., Verberne, S., Walsh, D., Fuhr, N., Quaresma, P., Gonçalves, T., Larsen, B., Balog, K., Macdonald, C., Cappellato, L., Ferro, N., Faculteit der Geesteswetenschappen, Language and Computation (ILLC, FNWI/FGw), and ILLC (FGw)
- Subjects
User profile ,Exploit ,Computer science ,05 social sciences ,020207 software engineering ,Information needs ,02 engineering and technology ,Recommender system ,Clef ,Ranking (information retrieval) ,World Wide Web ,Metadata ,0202 electrical engineering, electronic engineering, information engineering ,0509 other social sciences ,User interface ,050904 information & library sciences - Abstract
The Social Book Search (SBS) Lab investigates book search in scenarios where users search with more than just a query, and look for more than objective metadata. Real-world information needs are generally complex, yet almost all research focuses instead on either relatively simple search based on queries, or on profile-based recommendation. The goal is to research and develop techniques to support users in complex book search tasks. The SBS Lab has three tracks. The aim of the Suggestion Track is to develop test collections for evaluating ranking effectiveness of book retrieval and recommender systems. The aim of the Interactive Track is to develop user interfaces that support users through each stage during complex search tasks and to investigate how users exploit professional metadata and user-generated content. The Mining Track focuses on detecting and linking book titles in online book discussion forums, as well as detecting book search research in forum posts for automatic book recommendation.
- Published
- 2016
39. Overview of the SBS 2016 Suggestion Track
- Author
-
Koolen, M., Bogers, T., Kamps, J., Balog, K., Cappellato, L., Ferro, N., Macdonald, C., Faculteit der Geesteswetenschappen, ILLC (FGw), Cultural Heritage and Identity, and Language and Computation (ILLC, FNWI/FGw)
- Published
- 2016
40. Are Topically Diverse Documents Also Interesting?
- Author
-
Azarbonyad, H., Saan, F., Dehghani, M., Marx, M., Kamps, J., Mothe, J., Savoy, J., Pinel-Sauvagnat, K., Jones, G.J.F., SanJuan, E., Cappellato, L., Ferro, N., and Information and Language Processing Syst (IVI, FNWI)
- Subjects
Information retrieval ,Relation (database) ,Computer science ,media_common.quotation_subject ,Quality (business) ,Low correlation ,Positive correlation ,media_common ,Diversity (politics) - Abstract
Text interestingness is a measure of assessing the quality of documents from users’ perspective which shows their willingness to read a document. Different approaches are proposed for measuring the interestingness of texts. Most of these approaches suppose that interesting texts are also topically diverse and estimate interestingness using topical diversity. In this paper, we investigate the relation between interestingness and topical diversity. We do this on the Dutch and Canadian parliamentary proceedings. We apply an existing measure of interestingness, which is based on structural properties of the proceedings (eg, how much interaction there is between speakers in a debate). We then compute the correlation between this measure of interestingness and topical diversity.Our main findings are that in general there is a relatively low correlation between interestingness and topical diversity; that there are two extreme categories of documents: highly interesting, but hardly diverse (focused interesting documents) and highly diverse but not interesting documents. When we remove these two extreme types of documents there is a positive correlation between interestingness and diversity.
- Published
- 2015
- Full Text
- View/download PDF
41. A Comparative Study of Click Models for Web Search
- Author
-
Grotov, A., Chuklin, A., Markov, I., Stout, L., Xumara, F., de Rijke, M., Mothe, J., Savoy, J., Kamps, J., Pinel-Sauvagnat, K., Jones, G.J.F., SanJuan, E., Cappellato, L., Ferro, N., and Information and Language Processing Syst (IVI, FNWI)
- Subjects
Information retrieval ,Computer science ,business.industry ,Ranging ,02 engineering and technology ,Space (commercial competition) ,Range (mathematics) ,Search engine ,Software ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Relevance (information retrieval) ,Baseline (configuration management) ,business ,Dynamic Bayesian network - Abstract
Click models have become an essential tool for understanding user behavior on a search engine result page, running simulated experiments and predicting relevance. Dozens of click models have been proposed, all aiming to tackle problems stemming from the complexity of user behavior or of contemporary result pages. Many models have been evaluated using proprietary data, hence the results are hard to reproduce. The choice of baseline models is not always motivated and the fairness of such comparisons may be questioned. In this study, we perform a detailed analysis of all major click models for web search ranging from very simplistic to very complex. We employ a publicly available dataset, open-source software and a range of evaluation techniques, which makes our results both representative and reproducible. We also analyze the query space to show what type of queries each model can handle best.
- Published
- 2015
- Full Text
- View/download PDF
42. Overview of the CLEF 2015 Social Book Search Lab
- Author
-
Koolen, M., Bogers, T., Gäde, M., Hall, M., Huurdeman, H., Kamps, J., Skov, M., Toms, E., Walsh, D., Mothe, J., Savoy, J., Pinel-Sauvagnat, K., Jones, G.J.F., SanJuan, E., Cappellato, L., Ferro, N., ILLC (FGw), Language and Computation (ILLC, FNWI/FGw), Mothe, Josiane, Savoy, Jacques, Kamps, Jaap, Pinel-Sauvagnat, Karen, Jones, Gareth J.F., SanJuan, Eric, Cappellato, Linda, and Ferro, Nicola
- Subjects
Cognitive models of information retrieval ,020 Bibliotheks- und Informationswissenschaften ,evaluation ,Information retrieval ,Concept search ,Computer science ,Information needs ,Recommender system ,search engines ,Ranking (information retrieval) ,World Wide Web ,Metadata ,ddc:020 ,Human–computer information retrieval ,ddc:000 ,000 Informatik, Informationswissenschaft, allgemeine Werke ,information retrieval ,ddc:004 ,User interface ,004 Datenverarbeitung ,Informatik - Abstract
The Social Book Search (SBS) Lab investigates book search in scenarios where users search with more than just a query, and look for more than objective metadata. Real-world information needs are generally complex, yet almost all research focuses instead on either relatively simple search based on queries or recommendation based on profiles. The goal is to research and develop techniques to support users in complex book search tasks. The SBS Lab has two tracks. The aim of the Suggestion Track is to develop test collections for evaluating ranking effectiveness of book retrieval and recommender systems. The aim of the Interactive Track is to develop user interfaces that support users through each stage during complex search tasks and to investigate how users exploit professional metadata and user-generated content.
- Published
- 2015
- Full Text
- View/download PDF
43. The Value of Multistage Search Systems for Book Search
- Author
-
Huurdeman, H., Kamps, J., Koolen, M., Kumpulainen, S., Cappellato, L., Ferro, N., Jones, G.J.F., San Juan, E., Language and Computation (ILLC, FNWI/FGw), ILLC (FGw), Cultural Heritage and Identity, and Faculteit der Geesteswetenschappen
- Published
- 2015
44. Overview of the SBS 2015 Suggestion Track
- Author
-
Koolen, M., Bogers, T., Kamps, J., Cappellato, L., Ferro, N., Jones, G.J.F., San Juan, E., Language and Computation (ILLC, FNWI/FGw), ILLC (FGw), Cultural Heritage and Identity, and Faculteit der Geesteswetenschappen
- Published
- 2015
45. Overview of the SBS 2015 interactive track
- Author
-
Gäde, M., Hall, M., Huurdeman, H., Kamps, J., Koolen, M., Skov, M., Toms, E., Walsh, D., Cappellato, L., Ferro, N., Jones, G.J.F., San Juan, E., Language and Computation (ILLC, FNWI/FGw), ILLC (FGw), and Faculteit der Geesteswetenschappen
- Abstract
Users looking for books online are confronted with both professional meta-data and user-generated content. The goal of the Interactive Social Book Search Track was to investigate how users used these two sources of information, when looking for books in a leisure context. To this end participants recruited by four teams performed two different tasks using one of two book-search interfaces. Additionally one of the two interfaces also investigated whether user performance can be improved by providing a user-interface that supports multiple search stages.
- Published
- 2015
46. Extended Overview of the Living Labs for Information Retrieval Evaluation (LL4IR) CLEF Lab 2015
- Author
-
Schuth, A., Balog, K., Kelly, L., Cappellato, L., Ferro, N., Jones, G.J.F., San Juan, E., and Information and Language Processing Syst (IVI, FNWI)
- Published
- 2015
47. Overview of the CLEF eHealth Evaluation Lab 2016
- Author
-
Fuhr, N, Balog, K, Ferro, N, Larsen, B, Quaresma, P, Goncalves, T, Macdonald, C, Cappellato, L, Kelly, Liadh, Goeuriot, Lorraine, Suominen, Hanna, Neveol, Aurelie, Palotti, Joao, Zuccon, Guido, Fuhr, N, Balog, K, Ferro, N, Larsen, B, Quaresma, P, Goncalves, T, Macdonald, C, Cappellato, L, Kelly, Liadh, Goeuriot, Lorraine, Suominen, Hanna, Neveol, Aurelie, Palotti, Joao, and Zuccon, Guido
- Abstract
In this paper we provide an overview of the fourth edition of the CLEF eHealth evaluation lab. CLEF eHealth 2016 continues our evaluation resource building efforts around the easing and support of patients, their next-of-kins and clinical staff in understanding, accessing and authoring eHealth information in a multilingual setting. This year’s lab offered three tasks: Task 1 on handover information extraction related to Australian nursing shift changes, Task 2 on information extraction in French corpora, and Task 3 on multilingual patient-centred information retrieval considering query variations. In total 20 teams took part in these tasks (3 in Task 1, 7 in Task 2 and 10 in Task 3). Herein, we describe the resources created for these tasks, evaluation methodology adopted and provide a brief summary of participants to this year’s challenges and some results obtained. As in previous years, the organizers have made data and tools associated with the lab tasks available for future research and development.
- Published
- 2016
48. Assessors agreement: A case study across assessor type, payment levels, query variations and relevance dimensions
- Author
-
Fuhr, N, Balog, K, Ferro, N, Larsen, B, Quaresma, P, Goncalves, T, Macdonald, C, Cappellato, L, Palotti, Joao, Zuccon, Guido, Bernhardt, Johannes, Hanbury, Allan, Goeuriot, Lorraine, Fuhr, N, Balog, K, Ferro, N, Larsen, B, Quaresma, P, Goncalves, T, Macdonald, C, Cappellato, L, Palotti, Joao, Zuccon, Guido, Bernhardt, Johannes, Hanbury, Allan, and Goeuriot, Lorraine
- Abstract
Relevance assessments are the cornerstone of Information Retrieval evaluation. Yet, there is only limited understanding of how assessment disagreement influences the reliability of the evaluation in terms of systems rankings. In this paper we examine the role of assessor type (expert vs. layperson), payment levels (paid vs. unpaid), query variations and relevance dimensions (topicality and understandability) and their influence on system evaluation in the presence of disagreements across assessments obtained in the different settings. The analysis is carried out in the context of the CLEF 2015 eHealth Task 2 collection and shows that disagreements between assessors belonging to the same group have little impact on evaluation. It also shows, however, that assessment disagreement found across settings has major impact on evaluation when topical relevance is considered, while it has no impact when understandability assessments are considered.
- Published
- 2016
49. The IR Task at the CLEF eHealth evaluation lab 2016: User-centred health information retrieval
- Author
-
Balog, K, Ferro, N, Macdonald, C, Cappellato, L, Zuccon, Guido, Palotti, Joao, Goeuriot, Lorraine, Kelly, Liadh, Lupu, Mihai, Pecina, Pavel, Muller, Henning, Budaher, Julie, Deacon, Anthony, Balog, K, Ferro, N, Macdonald, C, Cappellato, L, Zuccon, Guido, Palotti, Joao, Goeuriot, Lorraine, Kelly, Liadh, Lupu, Mihai, Pecina, Pavel, Muller, Henning, Budaher, Julie, and Deacon, Anthony
- Published
- 2016
50. CLEF : Conference on Multilingual and Multimodal Information Access Evaluation : Working Notes for CLEF 2014 Conference Sheffield, UK, September 15-18, 2014
- Author
-
Cappellato, L., Ferro, N., Halvey, M., and Kraaij, W.
- Subjects
CEUR Workshop Proceedings ,Data Science - Abstract
Item does not contain fulltext 1558 p.
- Published
- 2014
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.