• Aucun résultat trouvé

Ideas for a Standard LL4IR Extension - Living Labs from a System Operator's Perspective

N/A
N/A
Protected

Academic year: 2022

Partager "Ideas for a Standard LL4IR Extension - Living Labs from a System Operator's Perspective"

Copied!
2
0
0

Texte intégral

(1)

Ideas for a Standard LL4IR Extension

Living Labs from a System Operator’s Perspective

Philipp Schaer1 and Narges Tavakolpoursaleh2

Cologne University of Applied Sciences, Cologne, Germany philipp.schaer@th-koeln.de

GESIS – Leibniz Institute for the Social Sciences, Cologne, Germany narges.tavakolpoursaleh@gesis.org

Abstract. We introduce the idea of developing a standard extension for common search engines and repository systems. This would not only increase the number of possible living labs participants on the site level but would additionally bring some other benefits like common standards and practices. We already developed such an extension for the repository system DSpace that might be a basis for future implementations.

1 Background and Problem Analysis

The idea of Living Labs for Information Retrieval has been successfully im- plemented in international IR evaluation campaigns like CLEF or TREC. To establish a robust and stable evaluation environment the LL4IR API is pub- licly available and is professionally hosted thanks to a funding by EFS ELIAS and Microsoft Azure. However, until today only five platforms implemented the API within their systems: REGIO J ´AT ´EK and Seznam[1], and the three aca- demic sites CiteseerX, Microsoft Academic Search, and the Social Science Open Access Repository SSOAR1. The latter being developed by the authors. While implementing the LL4IR component into SSOAR we learned: (1) The process of extracting head queries, compiling the JSON markup, establishing a work flow for uploading the feedback, implementing the interleaving, and many other tasks sum up and make it quite an effort to be part of LL4IR. (2) The query distri- bution of our system is highly skewed and not many typical head queries that are issued several hundred times a day are present. This might be due to the quite specific range of topics represented within the repository and the nature of academic search itself. Another reason might be that a lot of users are using the direct links provided by other search engines like Microsoft Academic Search.

2 A LL4IR Component for Popular Search Environments

SSOAR is DSpace-based repository system that uses a Solr search engine. While implementing the LL4IR component we paid attention to make it as minimally

1 http://trec-open-search.org/sites/

(2)

invasive and encapsulated as possible. This led to a quite reusable piece of soft- ware that might be used as an official extension for DSpace. We tested the ex- tension with the stable branches 3 and 5, both within an out-of-the-box vanilla installation and the specific implementation of SSOAR. We believe this to be a benefit for the whole repository community as this allows other repository op- erators to easily be part of the LL4IR community. There are more than 1,350 systems listed in OpenDOAR, a registry for Open Access Repositories, that are based on DSpace. A huge field of candidates for next year’s CLEF or TREC LL4IR campaigns. These different installations share a common system setup but featuring different content (e.g. repositories from the social sciences, arts or the sciences). This introduces the possibility to test ranking mechanisms in very different domains.

Due to many comparable systems within the same campaign it might be possible to surpass the missing head queries as the systems might share only their top n queries and sum up their head queries with other DSpace installations.

This wouldn’t lead to 100 head queries but maybe thousands. Why not use that many different queries and than later decide which of them are statistically stable enough to be included within the final evaluation round?

A standard LL4IR implementation introduces the possibilities to set some common standards and practices, like same timeout configurations, the guaran- tee that the interleaving algorithm is the same in every system, and so on.

By having both Microsoft Academic Search as a search engine and DSpace systems as the content-bearing repositories it might be possible to interlink search sessions. While a user is searching for scientific documents within Aca- demic Search he is later transferred to the repository where he can find the full text or additional document information. When both systems are part of the Living Labs campaign they might include a common URL parameter to indicate that this specific request is to be taken into account for the LL4IR campaign.

This way we can find out if users are interacting with the document like e.g.

bookmarking them, recommending them, tweeting about them or downloading the PDF file.

3 Conclusion and Outlook

All the things we listed above are still true for other popular search environments like Solr-based systems or content management systems like Typo3 or Wordpress.

We therefore would like to discuss whether the idea of using standard extension for popular search environments is worth the try and what other positive or negative outcomes might be possible.

References

1. Schuth, A., Balog, K., Kelly, L.: Overview of the living labs for information retrieval evaluation (ll4ir) clef lab 2015. In: CLEF 2015 - 6th Conference and Labs of the Evaluation Forum. Lecture Notes in Computer Science (LNCS), Springer (Septem- ber 2015)

Références

Documents relatifs

Our data warehouse query should provide the following features: (i) An intuitive usable graphical user interface to create easily queries, whose result is displayed in a clear way,

We compare the perceived usefulness of general descriptions, publication metadata, user reviews and tags in focused and open-ended search tasks, based on data gathered in the

Since entire patient admissions were mapped to timelines, after dates and times are extracted for all records in a patient admission, this metadata, along with the medical record

This extension com- bines document and process metadata in a visual represen- tation of search results to help people manage their infor- mation seeking.. We chose the browser

The perspective scores computed in section 2.1 are dis- played within the search results, and based on the perspec- tive score a document receives , we define four levels of

Complex information tasks shift the burden of information processing to al- gorithms that need to reason over heterogeneous data and perform constraint resolution, reasoning,

The search engine has been rigorously prototyped as part of the DARPA MEMEX program and has been integrated into the latest version of the Domain-specific Insight Graph

The language has been inspired by CQL, the language used for search- ing the Czech National Corpus, and uses similar key- words and syntax extended to facilitate searching in