Dynamic Transfer Learning for One-class Classification : a Multi-task Learning Approach
Texte intégral
Documents relatifs
First, it builds similarity matrices rather than clusters between rows and columns of a data matrix ; this is useful in the multi-view context since it allows us to combine easily
Dictionary learning has been widely used in reconstruction tasks. In its classical formulation, the goal is to learn a set of atoms directly from data. Using such unsupervised
As a consequence, we define a multi-task learning strategy using a hard parameter sharing neural network model that takes as input a learning word pair (x, y) encoded as
Evaluation of the Delorme and Thorpe model performances Regarding fast data classification, the Delorme and Thorpe model (Thorpe et al., 2001) is based on the fact that very
The proposed multi-task learning with dynamic weights is based on the deep CNNs (see Figure 1). The hard param- eter sharing structure is adopted as our framework, in which the
The contributions of this paper are threefold: i) we formalize a simple yet effective selective loss function that can be used in shared deep network to exploit training data
Since SCMTL forces all the tasks to have weights generated from a common distribution, it learns a very general prior, and thus, it cannot approximate the value functions as accurate
Although the Q' outputs could be represented by combinations of only the next largest integer to log~ IQ') binary outputs, it would require Q' binary perceptrons to learn the