• Aucun résultat trouvé

SafePredict: A Machine Learning Meta-Algorithm That Uses Refusals to Guarantee Correctness

N/A
N/A
Protected

Academic year: 2021

Partager "SafePredict: A Machine Learning Meta-Algorithm That Uses Refusals to Guarantee Correctness"

Copied!
2
0
0

Texte intégral

(1)

HAL Id: hal-02266300

https://hal.archives-ouvertes.fr/hal-02266300

Submitted on 13 Aug 2019

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

SafePredict: A Machine Learning Meta-Algorithm That

Uses Refusals to Guarantee Correctness

David Ramirez, Mustafa Kocak, Elza Erkip, Dennis Shasha

To cite this version:

(2)

Base Predictor

Confidence Based Refusal (CBR)

SafePredict

SafePredict + CBR

Amnesic SafePredict

SafePredict: A Machine Learning Meta-Algorithm

That Uses Refusals to Guarantee Correctness

David Ramirez

(dard@princeton.edu)

, Mustafa A. Kocak, Elza Erkip, and Dennis E. Shasha

Mustafa A. Kocak, D. Ramirez, et al., "SafePredict: A Meta-Algorithm for Machine Learning That Uses Refusals to Guarantee Correctness." Available on arXiv.

Nick Littlestone, and Manfred K. Warmuth, "The weighted majority algorithm." Information and computation, 1994.

Claudio De Stefano, et al., "To reject or not to reject: that is the question-an answer in case of neural classifiers." IEEE Trans. on Systems, Man, and Cybernetics, Part , 2000.

Li, Lihong, et al., "Knows what it knows: a framework for self-aware learning." Machine learning, 2011.

Amin Sayedi, et al., "Trading off mistakes and don't-know predictions." Advances in Neural Information Processing Systems, 2010.

Online prediction setup with refusal option.

Expert

Prediction

Meta-algorithm

with target

error rate

𝜖

Predict

𝑦

𝑃,𝑡

w. probability 𝑤

𝑃,𝑡

Refuse w. probability 1 − 𝑤

𝑃,𝑡

𝑦

𝑃,𝑡

𝑍

1

, … , 𝑍

𝑡−1

𝑋

𝑡

𝑤

𝑃,𝑡

Wolf

Beard

Man

Observation

Crowd of experts (i.e., algorithms) are asked to predict.

Dummy expert always refuses to predict.

Problem Setup

Introduction

References

Machine learning and prediction algorithms are the

building blocks of automation and forecasting.

Algorithms benefit from a lower error rate.

Definitions

𝑡 = time index, 𝑇 = total observations, 𝜂 = learning rate

𝑇

= σ

𝑡=1

𝑇

𝑤

𝑃,𝑡

, expected predictions

𝐿

𝑇

= σ

𝑡=1

𝑇

𝑙

𝑃,𝑡

𝑤

𝑃,𝑡

, expected cumulative loss

𝑉

= σ

𝑡=1

𝑇

𝑤

𝑃,𝑡

𝑤

𝐷,𝑡

, variance for number of predictions

𝑤

𝑃,𝑡+1

=

𝑤

𝑃,𝑡

𝑒

−𝜂𝑙𝑝,𝑡

𝑤

𝑃,𝑡

𝑒

−𝜂𝑙𝑝,𝑡

+𝑤

𝐷

𝑒

−𝜂𝜖

weight shift rule

Algorithm Properties

Def. A meta-algorithm is valid if, as 𝑇

→ ∞, average

expected loss ≤ target error rate.

Def. A meta-algorithm is efficient if, as 𝑇

→ ∞, refusals

occur only a finite number of times.

SafePredict, a meta-algorithm, takes predictions from underlying

algorithms and decides whether or not to predict with them.

Prediction ො

𝑦

𝑃,𝑡

or refusal ො

𝑦

𝐷

suffer a loss 𝑙

𝑃,𝑡

, 𝑙

𝐷

∈ [0,1].

Mistakes are costly, but we learn by observing.

Safe-Predict is valid and efficient!

Guaranteed with no assumptions on data or underlying experts, but asymptotic in the number of non-refused predictions.

Theorem 1.- With learning rate 𝜂 = Θ(

1

𝑉

), SafePredict is

guaranteed valid for any 𝑃. Particularly

𝐿

𝑇

𝑇

− 𝜖 = 𝑂

𝑉

𝑇

=

1

𝑇

.

Theorem 2.- If lim sup

𝑡→∞

𝐿

𝑃,𝑡

𝑡

< 𝜖 and

𝜂𝑇 → ∞, then SafePredict is efficient.

Experimental Results

Randomly permute data, choose first 10k points for experiment. Target error rate 𝜖 = 0.05.

Random label permutation introduced every 2k points (i.e., artificial change points).

Références

Documents relatifs

The goal of a meta-active learning system is to learn an active-learning strategy such that, for each problem, coming with a training dataset of unlabeled examples, it can predict

Like the face recognition and the natural language processing examples, most works discussing multi-task learning (Caruana 1997) construct ad-hoc combinations justified by a

Our approach to activity recognition and classification consists of two sequential tasks, pose estimation from images, and then the classification of the

Global-level meta-learners use the characteristics of data to select the overall- best algorithm for entire datasets, or an entire platform (e.g. Mid-level meta-learners select

Independently, each dataset is described using so-called meta-features. Meta- features are generally of two kinds: i) simple statistics of the dataset such as number of

Such rules can provide interesting insights on their own, but they are also very valuable features for more sophisticated schemes like Random Forests.. A hierarchical variant is able

The first observation we can make is that all the three variants of the sample-based active testing methods compete quite well with the (more complex) predecessor method (ATCV-ACC)

In our future work we would try to: use more rank aggregation algorithm models and algorithms; apply feature selection algorithms; introduce wider generic data class, and