• Aucun résultat trouvé

The University of Amsterdam at WePS3

N/A
N/A
Protected

Academic year: 2022

Partager "The University of Amsterdam at WePS3"

Copied!
1
0
0

Texte intégral

(1)

The University of Amsterdam at WePS3

Manos Tsagkias, Krisztian Balog

ISLA, University of Amsterdam, The Netherlands e.tsagkias@uva.nl,k.balog@uva.nl

Abstract. In this paper we describe our participation in the Third Web People Search (WePS3) evaluation campaign. We took part in the Online Reputation Management (ORM) task. Ambiguity of organization names (e.g., “Amazon” or

“Apple”) raises obvious difficulties for systems that attempt to trace mentions of and opinions about a specific company in Web data, in an unsupervised manner.

Problems are further amplified in the context of user generated content, where proper capitalization of named entities is often absent. The ORM task, introduced this year, addresses this very problem, by setting out the following challenge:

given a set of Twitter entries containing an (ambiguous) company name and given the homepage of the company, discriminate entries that do not refer the company.

Given the above definition, it is natural to formulate the problem as a binary classification task. Our focus was on building a general organization classifier that predicts, for each tweet, whether it is about a company. Our goal is to assess how a system without external aid from other sources (the company’s homepage, Wikipedia entry, etc.) can perform. We, therefore, focus on extracting features that are organization-independent and build on the characteristics of Twitter, such as noisy text, abbreviations and Twitter-specific language.

Specifically, we trained a J48 decision tree classifier using the following groups of features: (i) company name (matching based on character 3-grams), (ii) content value (whether the tweet contains URLs, hashtags or is part of a conversation), (iii) content quality (ratio of punctuation and capital characters), (iv) organiza- tional context (ratio of words found in tweets labelled as positive).

We submitted a single run that performed around the median of all submitted systems. One interesting observation that requires further investigation is that our F-score for the negative class was substantially higher than for the positive class (0.55vs.0.36); for other teams it was usually the other way around.

In future work we plan to build company-specific models by exploiting content both from Twitter and from external sources.

Acknowledgements This research was supported by the Netherlands Organisation for Scientific Research (NWO) under project number 612.061.815 and partially by the Cen- ter for Creation, Content and Technology (CCCT).

Références

Documents relatifs

We describe the effects of using character n-grams and field combinations on both monolingual English retrieval, and cross- lingual Dutch to English

We experimented with retrieval units (sentences and paragraphs) and with the similarity functions used for centrality computations (word overlap and cosine similarity).. We found

These four question processing streams use the two data sources for this task, the Dutch CLEF newspaper corpus and Dutch Wikipedia, as well as fact tables which were generated

We experimented with the combination of content and title runs, using standard combination methods as introduced by Fox and Shaw [1]: combMAX (take the maximal similarity score of

In order to make it possible for our answer re-ranking module (described in section 6) to rank answers from different streams, we took advantage of answer patterns from

Given a topic (and associated article) in one language, identify relevant snippets on the topic from other articles in the same language or even in other languages.. In the context

Language restricted Since we know the language of the topic in each of the translations, and the intention of the translated topic is to retrieve pages in that language, we may

When we examined the assessments of the answers to the three different question types, we noticed that the proportion of correct answers was the same for definition questions (45%)