• Aucun résultat trouvé

Online F-Measure Optimization

N/A
N/A
Protected

Academic year: 2022

Partager "Online F-Measure Optimization"

Copied!
1
0
0

Texte intégral

(1)

Online F-Measure Optimization

R´obert Busa-Fekete1, Bal´azs Sz¨or´enyi2, Krzysztof Dembczy´nski3, and Eyke H¨ullermeier1

1 Department of Computer Science, University of Paderborn, Warburger Str. 100, 33098 Paderborn, Germany

{busarobi,eyke}@upb.de

2 MTA-SZTE Research Group on Artificial Intelligence, Tisza Lajos krt. 103., H-6720 Szeged, Hungary

szorenyi@inf.u-szeged.hu

3 Institute of Computing Science, Pozna´n University of Technology, Piotrowo 2, 60-965 Pozna´n, PolandKrzysztof.Dembczynski@cs.put.poznan.pl

Abstract. 4 The F-measure is an important and commonly used per- formance metric for binary prediction tasks. By combining precision and recall into a single score, it avoids disadvantages of simple metrics like the error rate, especially in cases of imbalanced class distributions. The problem of optimizing the F-measure, that is, of developing learning algo- rithms that perform optimally in the sense of this measure, has recently been tackled by several authors. In this paper, we study the problem of F-measure maximization in the setting of online learning. We propose an efficient online algorithm and provide a formal analysis of its con- vergence properties. Moreover, first experimental results are presented, showing that our method performs well in practice.

Keywords: classification, learning theory, F-measure, structured out- put prediction

Copyright c 2015 by the paper’s authors. Copying permitted only for private and academic purposes. In: R. Bergmann, S. G¨org, G. M¨uller (Eds.): Proceedings of the LWA 2015 Workshops: KDML, FGWM, IR, and FGDB. Trier, Germany, 7.-9.

October 2015, published at http://ceur-ws.org

4 This work is accepted at the Twenty-ninth Annual Conference on Neural Information Processing Systems (NIPS).

106

Références

Documents relatifs

Figure 2: Expected overhead (a) and optimal length (b) of a pattern against the number of partial verifications when a single type of detector is used for three different

In this section, we present an experimental evaluation of our method on public and real private datasets with comparisons to classic distance-based methods and state of the art

We give a very simple algorithm to compute the error and complemen- tary error functions of complex argument to any given

a given measure µ on N , because this is the situation appearing in the problem of the calculation of the scaled entropy of the next-jump time filtrations.. For example, condition

Let M be a compact connected manifold of dimension n endowed with a conformal class C of Riemannian metrics of volume one.. Colbois : Universit´e de Neuchˆatel, Laboratoire

Although the algorithm crucially relies on the knowledge of the trapdoor R to perform this sampling efficiently, the distribution it outputs should be independent of R given the

A detailled analysis of the algorithms indicate that LIBSVM runs best when the cache contains all the dot products involving a potential support vector and an arbitrary example:

Here, we are concerned with the problem of maximization of expected power utility of both terminal wealth and consumption, in a market with investment opportunities in a savings