• Aucun résultat trouvé

Deep Saliency Models : the Quest for the Loss Function

N/A
N/A
Protected

Academic year: 2021

Partager "Deep Saliency Models : the Quest for the Loss Function"

Copied!
11
0
0

Texte intégral

(1)

HAL Id: hal-02264898

https://hal.inria.fr/hal-02264898

Preprint submitted on 7 Aug 2019

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Deep Saliency Models : the Quest for the Loss Function

Alexandre Bruckert, Hamed Tavakoli, Zhi Liu, Marc Christie, Olivier Le Meur

To cite this version:

Alexandre Bruckert, Hamed Tavakoli, Zhi Liu, Marc Christie, Olivier Le Meur. Deep Saliency Models :

the Quest for the Loss Function. 2019. �hal-02264898�

(2)

DEEP SALIENCY MODELS : THE QUEST FOR THE LOSS FUNCTION Alexandre Bruckert 1 , Hamed R. Tavakoli 2 , Zhi Liu 3 , Marc Christie 1 , Olivier Le Meur 1

1 Univ Rennes, IRISA, CNRS, France

2 Aalto University, Finland

3 Shangai University, China

ABSTRACT

Recent advances in deep learning have pushed the perfor- mances of visual saliency models way further than it has ever been. Numerous models in the literature present new ways to design neural networks, to arrange gaze pattern data, or to ex- tract as much high and low-level image features as possible in order to create the best saliency representation. However, one key part of a typical deep learning model is often neglected:

the choice of the loss function.

In this work, we explore some of the most popular loss func- tions that are used in deep saliency models. We demonstrate that on a fixed network architecture, modifying the loss func- tion can significantly improve (or depreciate) the results, hence emphasizing the importance of the choice of the loss function when designing a model. We also introduce new loss functions that have never been used for saliency prediction to our knowledge. And finally, we show that a linear combina- tion of several well-chosen loss functions leads to significant improvements in performances on different datasets as well as on a different network architecture, hence demonstrating the robustness of a combined metric.

1. INTRODUCTION

Despite decades of research, visual attention mechanisms of humans remain complex to understand and even more com- plex to model. With the availability of large databases of eye- tracking and mouse movements recorded on images [1, 2], there is now a far better understanding of the perceptual mech- anisms. Significant progress has been made in trying to pre- dict visual saliency, i.e. computing the topographic represen- tation of visual stimulus strengths across an image. Deep saliency models have strongly contributed to this progress.

However, as recently pointed out by Borji [3], a neglected challenge in the design of a deep saliency model is the choice of an appropriate loss function. In [4] a probabilistic end-to- end framework was proposed and five relevant loss functions were studied. Yet, to the best of our knowledge, none of the papers concerning the challenges in designing deep saliency models have investigated this aspect properly, despite its in- fluence on the quality of the results. Important questions

therefore arise: how do different loss functions affect per- formance of deep saliency networks? Which loss functions perform better than others and on which metrics? Is there ac- tually substantial benefits in combining loss functions? And how does the combination of loss functions perform with re- spect to individual loss functions? In this work, we seek an- swers to such questions by conducting a series of extensive experiments with both well-known and newly designed loss functions.

For this purpose, we first categorize loss functions per type of metric : (i) pixel-based comparisons e.g. Mean Square Error, Absolute Errors Exponential Absolute Difference, (ii) distribution-based metrics e.g. Kullback-Leibler diver- gence, Bhattacharya loss, binary cross-entropy, (iii) saliency- inspired metrics such as Normalized Scanpath Saliency or Pearson’s Correlation Coefficient, or (iv) perceptual-based metrics, gathering two novel metrics we propose in this paper which are inspired from image style transfer, and measure the aggregation of distances computed at each convolutional layer, between the convoluted reference image and the gener- ated saliency map.

We then design a novel deep saliency model to provide a fixed network architecture as a reference on which all the loss functions will be evaluated. Our evaluation strategy then consists in evaluating the impact of all the loss functions taken individually, on our fixed network with a fixed image dataset (MIT). Then by building on the common agreement that different metrics favor different perceptual characteristics of the image [3], we propose to further explore how the com- bination of loss functions, typically aggregating pixel-based, distribution-based, saliency-based and perceptual-based func- tions, can significantly influence the quality of the training.

To demonstrate the generalization capacity of our combined metric, we measure its impact on different datasets (CAT2000 and FiWi) and also with a different network architecture (SAM-VGG).

The contributions of this paper are therefore (i) to demonstrate

how the choice of the loss function can strongly improve (or

depreciate) the quality of a deep saliency model, and (ii) how

an aggregation of carefully selected loss functions can lead to

significant improvements, both on the fixed network architec-

ture we proposed, but also on some other architectures and

(3)

datasets.

The paper is organized as follows. Section 2 presents the re- lated works. The loss functions for training a deep architec- ture aiming to predict saliency map are described in Section 3.

Section 4 presents the comprehensive analysis of loss func- tions and their combinations. Conclusions are drawn in the last section.

2. RELATED WORKS

Computational models of saliency prediction, a long stand- ing problem in computer vision, have been studied from so many perspectives that going through all is beyond the scope of this manuscript. We, thus, provide a brief account of rele- vant works and summarize them in this section. We refer the readers to [5, 6] for an overview.

To date, from a computer vision perspective, we can divide the research on computational models of saliency prediction into two era (1) pre-deep learning, and (2) deep learning.

During the pre-deep learning period, significant number of saliency models were introduced, e.g. [7, 8, 9, 10, 11], and numerous survey papers looked into these models and their properties, e.g. [5, 12]. During this period the community converged into adopting eye tracking as a medium for ob- taining ground truth and dealt with challenges regarding the evaluation and the models, e.g. [13, 14]. This era was then replaced by saliency models based on deep learning tech- niques [3], which will be the main focus of this paper.

We therefore outline the recent research developments of deep saliency model era from two perspectives, (1) challenges of deep models and works that addressed them, and (2) the deep saliency models. We, then, stress the importance of task spe- cific loss functions in computer vision.

Challenges of deep saliency models.

The use of deep learning introduced new challenges to the community. The characteristics of most of the models shifted towards data intensive models based on deep convolutional neural networks (CNNs). To train a model, a huge amount of data is required; motivating the search for alternatives to eye tracking databases like mouse tracking [2], or pooling all the existing eye tracking databases into one [15].

To improve the training, Bruce et al. [15] investigated the factors required to take into account when relying on deep models, e.g., pre-processing steps, tricks for pooling all the eye tracking databases together and other nuances of training a deep model. Authors, however, considered only one loss function in their study.

Tavakoli et al. [16] looked into the correlation between mouse tracking and eye tracking at finer details, showing the data from the two modalities are not exactly the same. They demonstrated that, while mouse tracking is useful for training a deep model, it is less reliable for model selection and eval- uation in particular when the evaluation standards are based

on eye tracking.

Given the sudden boost in overall performance by saliency models using deep learning techniques, Bylinskii et al. [17]

reevaluated the existing benchmarks and looked into the fac- tors influencing the performance of models in a finer detail.

They quantified the remaining gap between models and hu- man. They argued that pushing performance further will re- quire high-level image understanding.

Recently Sen et al. [18] investigated the effect of model train- ing on neuron representations inside a deep saliency model.

They demonstrated that (1) some visual regions are more salient than others, and (2) the change in inner-representations is due to the task that original model is trained on prior to be- ing fine-tuned for saliency.

Deep saliency models.

The deep saliency models fall into two categories, (1) those using CNNs as a fixed feature extractors and learn a regres- sion from feature space into saliency space using a none- neural technique, and (2) those that train a deep saliency model end-to-end. The number of models belonging to the first category is limited. They are not comparable within the context of this research because the regression is often car- ried out such that the error can not be back-propagated, e.g., [19] employs support vector machines and [20] uses extreme learning machines. Our focus is, however, the second group.

Within end-to-end deep learning techniques, the main re- search has been on architecture design. Many of the models borrow the pre-trained weights of an image recognition net- work and experiment combining different layers in various ways. In other words, they engineer an encoder-decoder net- work that combines a selected set of features from different layers of a recognition network. In the following we discuss some of the most well-known models.

Huang et al. [21] proposed a multi-scale encoder based on VGG networks and learns a linear combination from re- sponses of two scales (fine and coarse). K¨ummerer et al. [22]

use a single scale model using features from multiple layers of AlexNet. Similarly, K¨ummerer et al. [23] and Cornia et al. [24] employed single scale models with features from multiple layers of a VGG architecture.

There has been also a wave of models incorporating recur- rent neural architectures. Han and Liu [25] proposed a multi- scale architecture using convolutional long-short-term mem- ory (ConvLSTM). It is followed by [26] using a slight mod- ified architecture using multiple layers in the encoder and a different loss function. Recurrent models of saliency predic- tion are more complex than feed-forward models and more difficult to train. Moreover, their performance is not yet sig- nificantly better than some recent feed-forward networks such as EML-NET [27].

In the literature of deep saliency models, a loss function or

a combination of several ones is chosen based on intuition,

expertise of the authors or sometimes mathematical formu-

(4)

lation of a model. K¨ummerer et al. [28] introduces the idea that information-theory can be a good inspiration for saliency metrics. They use the information gain to explain how well a model performs compared to a gold-standard baseline.

Consequently, they use the log-likelihood for a loss function in [29], achieving state-of-the-art results in saliency predic- tion. Jetley et al. [4] are part of the very few who specifically focused on the design of a loss functions for saliency mod- els. They proposed the use of Bhattacharyya distance and compared it to 4 other probability distances. In this paper, in contrast to [4], we (1) adopt a principled approach to compare existing loss functions and their combinations and (2) inves- tigate their convergence properties over different datasets and network architectures.

Deep Learning, loss functions and computer vision.

With the application of deep learning techniques to computer vision domain, the choice of appropriate loss function for a task has become a critical aspect of the model training. The computer vision community have been successful in develop- ing task tailored loss functions to improve a model, e.g., en- coding various geometric properties for pose estimation [30], curating loss functions enforcing perceptual properties of vi- sion for various generative deep models [31], exploiting the sparsity within the structure of problem, e.g., class imbal- anced between background and foreground in detection prob- lem, for reshaping standard loss functions and form a new effective loss functions [32]. Our efforts follows the same path to identify the effectiveness of a range of loss functions in saliency prediction.

3. LOSS FUNCTIONS FOR DEEP SALIENCY NETWORK

Before delving into the description of loss functions, we present the architecture of the convolutional neural network that will be used throughout this paper. After this presenta- tion, we elaborate on the tested loss functions.

3.1. Proposed baseline architecture

Figure 1 presents the overall architecture of the proposed model. The purpose of designing a new architecture is only to perform a comparison with existing architectures. Our architecture is based on the deep gaze network of [29] and on the multi-level deep network of [24]. The pre-trained VGG-16 network [33] is used for extracting deep features of an input image (400 × 300) from layers conv3 pool, conv4 pool, conv5 conv3. Feature maps of layers conv4 pool and conv5 conv3 are rescaled to get feature maps with a sim- ilar spatial resolution.

That feature map with 1280 channels is then fed into a shallow network composed of the following layers: a first convolu- tional layer allows us to reduce by a factor ten the number of

channels, which are then processed by an ASPP (an atrous spatial pyramid pooling [34]) of 4 levels. Each level has a convolution kernel of 3 × 3, a stride equal to 1 and a depth of 32. The dilatation rates are 1, 3, 6, and 12. The ASPP benefit is to catch information in a coarse-to-fine approach while keeping the resolution of the input feature maps. The output of the four pyramid levels are then merged together, i.e. leading to 4 ×32 maps. The last 1 × 1 convolutional layer reduces the data dimensionality to 1 feature map. This map is then smoothed by a Gaussian filter 5 × 5 with a standard deviation of 1. The activation function of these layers is a ReLU activation.

The network was trained over the MIT dataset composed of more than 1000 images [35]. We split this dataset into 500 images for the training, 200 images for the validation and the rest for the test. We use a batch size of 60, and the stochastic gradient descent. To prevent over-fitting, a dropout layer, with a rate of 0.25, is added on top the network. The learning rate is set to 0.001. During the training, the network was val- idated against the validation test to monitor the convergence and to prevent over-fitting. The number of trainable parame- ters is approximately 1,62 millions. In the following section, we present the different tested loss functions used during the training phase.

Fig. 1: Architecture of the proposed deep network.

3.2. Loss functions

Let I : Ω ⊂ R 2 7→ R 3 an input image of resolution N × M .

We suppose S and S ˆ the vectorized human and the predicted

saliency maps, i.e. S and S ˆ are in R N×M . Let also S f ix be

the human eye fixations map, i.e. a N × M image with 1 or

0 pixels. In the following, we present thirteen loss functions

L tested in this study. There are classified into four categories

according to their characteristics: pixel-based, distribution-

based, saliency-inspired and perceptual-based.

(5)

3.2.1. Pixel-based loss functions

For pixel-based loss functions L, we assume that S and S ˆ are in [0, 1]. We evaluate the following loss functions:

• Mean Squared Error (MSE) measures the averaged squared error between prediction and ground truth:

L(S, S) = ˆ 1 N × M

N ×M

X

i=1

(S i − S ˆ i ) 2 (1)

• Exponential Absolute difference (EAD):

L(S, S) = ˆ 1 N × M

N ×M

X

i=1

exp(|S i − S ˆ i |) − 1 (2)

• Absolute Error (AE):

L(S, S) = ˆ 1 N × M

N ×M

X

i=1

|S i − S ˆ i | (3)

• Weighted MSE Loss (W-MSE):

L(S, S) = ˆ 1 N × M

N×M

X

i=1

w i

S i − S ˆ i

2

(4)

The weight w i allows to put more emphasis on errors occurring on salient pixels of the ground truth S. Two functions are tested in this paper. In [24], authors defined the loss function (MLNET): w i = α−S 1

i

and α = 1.1.

Therefore, when S i = 1, the error is multiplied by a factor 10, whereas when S i = 0, the multiplying factor is equal to 0.90. We also consider a weighting func- tion based on a parametric sigmoid function (SIG-MSE):

w i = 1+exp(−k×(S k

i

−λ)) , where k = 10 and λ varies between 0 and 1.

3.2.2. Distribution-based loss functions

For the distribution-based loss functions, we consider that the vectorized human and the predicted saliency pixels represent a probability to be salient. For that the network presented in section 3.1 is modified in order to output pixel-wise predic- tions that can be considered as probabilities for independent binary random variables [36]. An element-wise sigmoid ac- tivation function is then added as being the last layer. The following loss functions are investigated:

• Kullback-Leibler divergence (KLD) measures the diver- gence between the distribution S and S: ˆ

L(S, S ˆ ) =

N ×M

X

i=1

S ˆ i log S ˆ i

S i

(5)

• Bhattacharya loss (BHAT) measures the similarity between the distribution S and S: ˆ

L(S, S) = ˆ

N×M

X

i=1

q

S i S ˆ i (6)

• Binary Cross Entropy (BCE) assumes that the saliency pre- diction S ˆ as well as the ground truth saliency map S are composed of independent binary random variables:

L(S, S) = ˆ −

N×M

X

i=1

S i log ˆ S i + (1 − S i ) log(1 − S ˆ i )

(7)

• Weighted Binary Cross Entropy (W-BCE): compared to the BCE loss, a global weight w is introduced to consider that there are much more non salient areas than salient ar- eas [37]. It allows to put more emphasis on errors occurring when S → 1 and S ˆ → 0 (w >> 0.5) or when S → 0 and S ˆ → 1 (w << 0.5):

L(S, S) ˆ = −

N ×M

X

i=1

w × S i log ˆ S i

+(1 − w)(1 − S i ) log(1 − S ˆ i ) (8)

• Focal Loss (FL) : In order to deal with the large foreground- background class imbalance encountered during the train- ing of dense detectors, Linet al. [32] modified the binary cross entropy loss function. Such class imbalance is also relevant in the context of saliency prediction, for which the ground truth saliency map mainly consists of null or close to zero, creating a similar phenomenon. The approach is quite similar to W-BCE, except that the weight is locally adjusted and based on a tunable γ power of the predicted saliency. As in [32], we set by default the γ value equal to 2:

L(S, S) ˆ = −

N×M

X

i=1

(1 − S ˆ i γ ) × S i log ˆ S i

+ ˆ S i γ (1 − S i ) log(1 − S ˆ i )

(9)

• Negative Logarithmic Likelihood (NLL): As shown by K¨ummerer et al. [28], information theory provides strong insights when it comes to saliency models. For instance, in [29], they use maximum likelihood learning. Let I be the set of fixations in an image, and N i the number of those fix- ations. The logarithm of the prediction at the coordinates of each fixation is then computed :

L( ˆ S) = − 1 N i

X

i∈I

log ˆ S i

(10)

(6)

3.2.3. Saliency-inspired loss functions

Saliency predictions are usually evaluated using several met- rics [38]. Those metrics are good candidates to use as loss functions, since they capture several properties that are spe- cific to saliency maps.

• Normalized Scanpath Saliency (NSS): This metric was in- troduced in [39], to evaluate the degree of congruency be- tween human eye fixations and a predicted saliency map.

Instead of relying on a saliency map as ground truth, the predictions are evaluated against the true fixations map.

The value of the saliency map at each fixation point is nor- malized with the whole saliency map variance:

L(S f ix , S) = ˆ 1 N × M

N ×M

X

i=1

S ˆ i − µ( ˆ S i )

σ( ˆ S i ) S i f ix (11)

• Pearson’s Correlation Coefficient (CC) measures the linear correlation between the ground truth saliency map and the predicted saliency map :

L(S, S) = ˆ σ(S, S) ˆ

σ(S)σ( ˆ S) (12) where σ(S, S) ˆ is the covariance of S and S. ˆ

3.2.4. Perceptual-based loss functions

We propose two new loss functions for deep saliency, that have been applied with success in image style-transfer prob- lems [31, 40]. The idea is to compare the representations of the ground truth and predicted saliency maps that are ex- tracted from different layers of a fixed pre-trained convolu- tional neural network. The idea behind those losses is to take into account not only the saliency map, but also the deep hid- den patterns that could exist, as well as the potential relation- ship between such patterns. Let φ j (S) be the activation at the j th layer of the VGG network when fed a saliency map S.

φ j (S) is then of size C j × H j × W j , where C j represents the number of filters, H j and W j represent the height and width of the feature maps at the layer j, respectively. We also denote J the set of layers from which we extract the representations.

In this work, we extracted the outputs of the 5 pooling layers of a fixed VGG-16 network [33] pre-trained on the ImageNet dataset, representing a total of 1920 filters:

• Deep Features loss (DF) measures the Euclidean distance between the feature representations:

L(S, S) = ˆ X

j∈J

1 C j × H j × W j

kφ j (S) − φ j ( ˆ S)k 2 (13)

• Gram Matrices of Deep Features loss (GM) : In order to leverage the potential statistical dependency between fea- tures maps, we propose a new loss relying on Gram matri- ces. For this purpose, we reshape the output φ j (S) into a

matrix ψ of size C j × (H j W j ). Then, for each layer j, the Gram matrix G φ j (S) of size C j × C j is defined as follows:

G φ j (S) = 1 C j × H j × W j

ψψ > (14)

The loss function is then the sum of the squared Frobenius norm of the difference between the Gram matrices G φ j , j ∈ J:

L(S, S) = ˆ X

j∈J

kG φ j (S) − G φ j ( ˆ S)k 2 F (15)

3.2.5. Center-bias regularization

Since our model does not take into account the center-bias with a learned prior, like in [26, 24], we add a regularization term. We compute a center-bias map B as the mean of the ground truth maps from the training part of the MIT dataset, and add to the loss function the regularization term R i for each pixel i:

R i = α( ˆ S i − B i ) 2 (16) We empirically set the parameter α to 0.1, even though it could be optimized to improve final results. This regulariza- tion will later be referred as R in Table 2. The center-bias map B is illustrated in Figure 2.

Fig. 2: Averaged colored saliency map of the training part of MIT dataset. Horizontal and vertical marginal distributions are also plotted, illustrating the center bias.

3.2.6. Linear combinations

All presented loss functions evaluate different characteristics

of the predicted saliency maps. We can then hypothesize that

a linear combination of some of those loss functions could

lead to better results, as it would aggregate the particularities

of all measures (a strategy already adopted by [26]). We de-

cided to evaluate two linear combinations: the first one (LC

(7)

1) combining KLD, CC and NSS loss functions, and the sec- ond one (LC 2) adding Deep Features loss, Gram Matrices loss and sigmoid-weighted MSE. This specific combination was chosen because it relies on an existing successful combi- nation and also aggregates the four types of metrics together.

We followed the work of [26] to set the coefficients for the first linear combination: −1 for the NSS, −2 for the CC, and 10 for the KLD. In the LC 2 combination, we kept those co- efficients and set arbitrarily all the other ones as 1. We used the best λ parameter that we tested for the sigmoid-weighted MSE in the linear combination (λ = 0.55).

4. EXPERIMENTS 4.1. Testing protocols

To carry out the evaluation, we use seven quality metrics ap- plied on the MIT benchmark [1, 38]: CC (correlation co- efficient, CC ∈ [−1, 1]), SIM (similarity, intersection be- tween histograms of saliency, SIM ∈ [0, 1]), AUC (Area Under Curve, AU C ∈ [0, 1]), NSS (Normalized Scanpath Saliency, N SS ∈ ]−∞, +∞[), EMD (Earth Mover Distance, EM D ∈ [0, +∞[) and KL (Kullback Leibler divergence, KL ∈ [0, +∞[).

The similarity degree between prediction and ground truth is computed over 299 images.

4.2. Performance of the proposed model

Table 1 presents the performance of the proposed model (when trained with the MSE loss function) compared to exist- ing models, i.e. Itti [7], Rare2012 [13], GBVS [9], AWS [41], Sam-ResNet & Sam-VGG [26], SalGan [36], ShallowNet

& DeepConvNet [42]. All models are evaluated on the test dataset as defined in Section 3.1.

Table 1: Performance of our model. Best performances are in bold. (AUC-B=AUC-Borji; AUC-J=AUC-Judd)

CC ↑ SIM ↑ AUC-J ↑ AUC-B ↑ NSS ↑

Itti 0.28 0.35 0.71 0.71 0.79

Rare2012 0.46 0.43 0.78 0.79 1.40

GBVS 0.49 0.44 0.81 0.81 1.38

AWS 0.39 0.40 0.75 0.76 1.22

Sam-ResNet 0.68 0.60 0.85 0.79 2.43

SalGan 0.70 0.58 0.86 0.86 2.17

Sam-VGG 0.64 0.57 0.85 0.78 2.19

ShallowNet 0.59 0.47 0.83 0.84 1.62

DeepConvNet 0.60 0.49 0.84 0.85 1.73

Our model 0.64 0.45 0.81 0.84 2.04

(ranking) 3/10 5/10 6/10 3/10 4/10

According to the evaluation, the proposed model performs rather well and is in the top 4 models. The best performing models are Sam-ResNet and SalGan. Do note however that

other models are not specifically trained on the MIT dataset, unlike ours.

4.3. Loss function performance

Table 2 presents the performance obtained with the different loss functions.

Which category of loss functions provide the best perfor- mances?

Except the perceptual-based loss functions, results suggest that the pixel-based, the distribution-based and the saliency- inspired loss functions perform similarly. However, the perceptual-based loss functions we introduced, namely DF and GM, do not perform well individually compared to the aforementioned losses. It might be due to the feature maps used for these losses which are extracted using a deep convo- lutional network that was trained using natural images, and not saliency maps. The representation of the saliency maps in the feature space might then not be appropriate for this task.

Designing a stronger loss from weaker losses.

Results also suggest that a simple linear combination of well known loss functions increases the ability of the network to predict saliency maps. While keeping the number of trainable parameters unchanged, we succeed in improving up to 14%

the correlation coefficient when we compared the best linear combination (CC = 0.7291) to the classical MSE (CC = 0.6388). In a more general way, linear combinations of the loss functions systematically improve the results on most of the metrics. Such fluctuations between the performances of the different loss functions confirm our hypothesis that the choice of the loss function is a critical part of designing a deep saliency model. Moreover, the aggregated loss of KL + CC + NSS + DF + GM improves the SIM, AUC-B and KL scores compared to the aggregation KL + CC + NSS. This reveals the influence of perceptual-based losses (DF+GM) in deep saliency models, and probably calls for future work in this direction.

Beyond this quantitative analysis, Figure 3 illustrates pre- dicted saliency maps obtained for some of the tested loss functions. Qualitatively speaking, the saliency maps obtained when using the combined loss KLD + CC + NSS + DF + GM + SIG-MSE + R look very similar to the ground truth maps. For instance, they are very condensed around the salient regions, with little noise.

4.4. Does the best loss generalize well over different datasets and a different architecture?

In this section, we test how well the best loss generalizes over two datasets, i.e. CAT2000 and FiWi, and one architecture, i.e. SAM-VGG.

CAT2000 and FiWi datasets.

(8)

Fig. 3: (a) Visual stimulus; (b) Ground truth saliency map; (c) KLD + CC + NSS + DF + GM + SIG-MSE + R combination;

(d) NSS; (e) KL + CC + NSS + R; (f) MLNET-MSE; (g) KLD; (h) SIG-MSE (λ = 0.55); (i) CC.

CAT2000 eye tracking dataset [43] is composed of 2000 images belonging to 20 different categories whereas FiWi dataset [44] is composed of more than 140 screen shots of webpages. Performances are given in Table 3. Results indi- cate that the loss function based on the linear combination of KLD, CC, NSS, DF, GM, SGI-MSE and R allows to signif- icantly increase the ability to predict visual saliency. Com- pared to the MLNET-MSE loss function, the gain in terms of CC is 16.3% and 7.1% for CAT2000 and FiWi datasets, respectively.

SAM-VGG with linear combinations of loss functions.

We also retrain SAM-VGG network over the training dataset, as described in Section 3.1, by considering two linear combi- naisons (LC 1 and LC 2). Table 4 presents the results.

The results confirm that the linear combination approach im- proves the prediction and hence generalizes well indepen- dently of the dataset and the network architecture. Even if the performances on the test datasets do not reach state-of-the- art techniques (see for instance [45] for FiWi), due to the fact that our model was only trained on natural images in MIT, the hierarchy of the loss functions we tested remains consistent, emphasizing the benefit of the linear combination. Figure 4 illustrates a predicted saliency map generated when the SAM- VGG network is trained with a stand-alone loss function, i.e.

MLNET-MSE, and with a combination of loss functions.

5. CONCLUSION

In this paper, we introduced a deep neural network which pur- pose was to evaluate the impact of loss functions on the pre-

Fig. 4: Example of good predictions by the combination loss while a single loss makes bad predictions (for SAM- VGG model). (a) original image; (b) Ground truth saliency map; (c) KLD + CC + NSS + DF + GM + SIG-MSE + R combination (CC = 0.8681 and 0.7967); (d) MLNET-MSE (CC = 0.4320 and 0.4491) .

diction capacity when it comes to deep saliency models. We evaluated several well-known and commonly used losses, and introduced a new kind of loss function (a perceptual-based loss) that, to the best of our knowledge, has not been applied to saliency prediction. These new loss functions seem to im- prove, at least partially, the performances of a deep saliency model. Further work on the exact contribution of this kind of losses however still remains necessary. We showed that a simple linear combination of different losses can signifi- cantly improve over individual losses, especially when when different types of loss functions are combined (pixel-based, distribution-based, perception-based, saliency-based).

More importantly we showed that this combination strategy

generalizes well on different datasets and also with a differ-

ent deep network architecture. Optimization on the coeffi-

(9)

Table 2: Performance of the loss functions. Best performances are in bold and the second and third best performances are in italic. R represents the center-biais regularization. (AUC-B=AUC-Borji; AUC-J=AUC-Judd)

CC ↑ SIM ↑ AUC-J ↑ AUC-B ↑ NSS ↑ EMD ↓ KL ↓ Pixel-based loss functions

MSE 0.6388 0.4492 0.8118 0.8363 2.0580 2.3279 0.9472

EAD 0.6725 0.4790 0.8326 0.8428 2.2133 2.1744 0.8592

AE 0.6426 0.4443 0.8145 0.8322 2.1388 2.3782 0.9616

MLNET-MSE 0.6904 0.5962 0.8416 0.8245 2.2468 1.3581 1.6042

SIG-MSE (λ = 0.25) 0.6929 0.5896 0.8512 0.8438 2.1897 1.4120 0.9876

SIG-MSE (λ = 0.55) 0.6725 0.5646 0.8505 0.8542 2.0799 1.5137 0.8478

SIG-MSE (λ = 0.75) 0.6440 0.5280 0.8512 0.8637 1.9480 1.7100 0.7686

Distribution-based loss functions

BCE 0.6616 0.4712 0.8231 0.8380 2.1229 2.2600 0.8899

W-BCE w = 0.9 0.6333 0.4287 0.8308 0.8519 1.8734 2.4793 1.0003

W-BCE w = 0.8 0.6363 0.4273 0.8305 0.8531 1.8909 2.5026 1.0067

W-BCE w = 0.7 0.6409 0.4308 0.8179 0.8396 1.9636 2.4824 0.9976

W-BCE w = 0.6 0.6478 0.4335 0.8182 0.8412 2.0135 2.4800 0.9862

W-BCE w = 0.5 0.6739 0.4305 0.8420 0.8582 2.0911 2.4713 0.9871

W-BCE w = 0.4 0.6443 0.3992 0.8301 0.8468 2.0166 2.6625 1.0949

Focal Loss 0.6530 0.4294 0.8197 0.8403 1.9552 2.4839 0.9738

KLD 0.6326 0.4893 0.8356 0.8541 1.7913 2.0609 0.8336

Bhat 0.6203 0.5029 0.8429 0.8567 1.7321 1.9209 0.7909

NLL 0.6251 0.4973 0.8407 0.8559 1.7856 1.8734 0.7955

Saliency-inspired loss functions

CC 0.6943 0.4994 0.8411 0.8386 1.8201 2.1378 0.9157

NSS 0.6740 0.4325 0.8397 0.8216 2.3142 2.9964 1.3498

Perceptual-based loss functions

Deep Features (DF) 0.6065 0.4772 0.8308 0.8259 1.9731 3.7546 0.9675

Gram Matrices (GM) 0.5911 0.4964 0.8371 0.8312 1.8357 2.0455 1.1993

Linear combinations

SIG-MSE + R 0.6813 0.5611 0.8507 0.8373 1.9734 3.1471 0.8349

KLD + CC + NSS 0.7288 0.5754 0.8512 0.8487 2.2464 2.1340 0.9571

KLD + CC + NSS + DF + GM 0.7192 0.5790 0.8492 0.8536 1.9652 2.3101 0.9387

KLD + CC + NSS + R 0.7176 0.5683 0.8579 0.8520 2.2147 2.8808 0.8912

KLD + CC + NSS + DF + GM + SIG-MSE + R 0.7291 0.5817 0.8585 0.8563 2.2094 2.5517 0.8010

Table 3: Performance of proposed model over CAT2000 and FiWi datasets with MLNET-MSE (W-MSE), KLD + CC + NSS + R (LC 1) and KLD + CC + NSS +DF +GM + SIG- MSE (λ = 0.55) + R (LC 2). Best performances are in bold.

(AUC-B=AUC-Borji; AUC-J=AUC-Judd)

CC ↑ SIM ↑ AUC-J ↑ AUC-B ↑ NSS ↑ CAT2000

W-MSE 0.508 0.4017 0.8221 0.8016 1.9486 LC 1 0.5535 0.4261 0.8273 0.8187 1.9332 LC 2 0.5937 0.4203 0.8309 0.8372 1.9375 FiWi

W-MSE 0.3954 0.3872 0.7312 0.7114 0.8050 LC 1 0.4118 0.4157 0.7621 0.7390 0.8214 LC 2 0.4236 0.4183 0.7636 0.7681 0.8205

Table 4: Performance of SAM-VGG over MIT dataset with MLNET-MSE (W-MSE), KLD + CC + NSS + R (LC 1) and KLD + CC + NSS +DF +GM + SIG-MSE + R (LC 2).

Best performances are in bold. (AUC-B=AUC-Borji; AUC- J=AUC-Judd)

CC ↑ SIM ↑ AUC-J ↑ AUC-B ↑ NSS ↑ MIT

W-MSE 0.7351 0.6769 0.8521 0.7884 2.0037

LC 1 0.7499 0.6502 0.8635 0.7912 2.1694

LC 2 0.7511 0.6472 0.8712 0.8017 2.0741

(10)

cients of those linear combinations is also possible to obtain the best performances possible out of the combination. This approach could moreover easily be extended to other kinds of architectures, not necessarily based on convolutional neural networks.

Finally, one of the main idea that motivated our work was to highlight the importance of the choice of the loss function.

We showed that a careful design of the loss function can sig- nificantly improve the performances of a model without in- creasing the number of trainable parameters.

6. REFERENCES

[1] Zoya Bylinskii, Tilke Judd, Ali Borji, Laurent Itti, Fr´edo Durand, Aude Oliva, and Antonio Torralba, “Mit saliency benchmark,” 2015.

[2] M. Jiang, S. Huang, J. Duan, and Q. Zhao, “Salicon:

Saliency in context,” in 2015 IEEE Conference on Com- puter Vision and Pattern Recognition (CVPR), 2015.

[3] Ali Borji, “Saliency prediction in the deep learn- ing era: An empirical investigation,” arXiv preprint arXiv:1810.03716, 2018.

[4] S. Jetley, N. Murray, and E. Vig, “End-to-end saliency mapping via probability distribution prediction,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.

[5] A. Borji and L. Itti, “State-of-the-art in visual attention modeling,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 1, pp. 185–207, 2013.

[6] A. Borji, D. N. Sihite, and L. Itti, “Quantitative analy- sis of human-model agreement in visual saliency model- ing: A comparative study,” IEEE Transactions on Image Processing, vol. 22, no. 1, pp. 55–69, Jan 2013.

[7] L. Itti, C. Koch, and E. Niebur, “A model of saliency- based visual attention for rapid scene analysis,” IEEE Transactions on Pattern Analysis and Machine Intelli- gence, vol. 20, no. 11, pp. 1254–1259, 1998.

[8] Neil D. B. Bruce and John K. Tsotsos, “Saliency based on information maximization,” in Proceedings of the 18th International Conference on Neural Information Processing Systems, 2005.

[9] Jonathan Harel, Christof Koch, and Pietro Perona,

“Graph-based visual saliency,” in Proceedings of the 19th International Conference on Neural Information Processing Systems, 2006.

[10] X. Hou, J. Harel, and C. Koch, “Image signature: High- lighting sparse salient regions,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no.

1, pp. 194–201, 2012.

[11] J. Zhang and S. Sclaroff, “Saliency detection: A boolean map approach,” in 2013 IEEE International Conference on Computer Vision, 2013.

[12] Qi Zhao and Christof Koch, “Learning saliency-based visual attention: A review,” Signal Processing, vol. 93, no. 6, pp. 1401–1407, 2013.

[13] Nicolas Riche, Matthieu Duvinage, Matei Mancas, Bernard Gosselin, and Thierry Dutoit, “Saliency and human fixations: State-of-the-art and study of compari- son metrics,” in The IEEE International Conference on Computer Vision (ICCV), 2013.

[14] A. Borji, H. R. Tavakoli, D. N. Sihite, and L. Itti, “Anal- ysis of scores, datasets, and models in visual saliency prediction,” in 2013 IEEE International Conference on Computer Vision, 2013.

[15] N. D. B. Bruce, C. Catton, and S. Janjic, “A deeper look at saliency: Feature contrast, semantics, and beyond,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 516–524.

[16] H. R. Tavakoli, F. Ahmed, A. Borji, and J. Laakso- nen, “Saliency revisited: Analysis of mouse movements versus fixations,” in 2017 IEEE Conference on Com- puter Vision and Pattern Recognition (CVPR), 2017, pp.

6354–6362.

[17] Z. Bylinskii, A. Recasens, A. Borji, A. Oliva, A. Tor- ralba, and F. Durand, “Where should saliency models look next?,” in European Conference on Computer Vi- sion (ECCV), 2016.

[18] Sen He, Hamed R Tavakoli, Ali Borji, Yang Mi, and Nicolas Pugeault, “Understanding and visual- izing deep visual saliency models,” arXiv preprint arXiv:1903.02501, 2019.

[19] E. Vig, M. Dorr, and D. Cox, “Large-scale optimization of hierarchical features for saliency prediction in natural images,” in IEEE Computer Vision and Pattern Recog- nition (CVPR), 2014.

[20] Hamed R. Tavakoli, Ali Borji, Jorma Laaksonen, and Esa Rahtu, “Exploiting inter-image similarity and en- semble of extreme learners for fixation prediction using deep features,” Neurocomput., vol. 244, no. C, pp. 10–

18, June 2017.

[21] Xun Huang, Chengyao Shen, Xavier Boix, and Qi Zhao,

“Salicon: Reducing the semantic gap in saliency predic-

tion by adapting deep neural networks,” in The IEEE

International Conference on Computer Vision (ICCV),

2015.

(11)

[22] M. Kummerer, L. Theis, and M. Bethge, “Deep gaze i:

Boosting saliency prediction with feature maps trained on imagenet,” in ICLR Workshop, 2015.

[23] M. Kummerer, T. S. Wallis, L. A. Gatys, and M. Bethge,

“Understanding low- and high-level contributions to fix- ation prediction,” in The IEEE International Conference on Computer Vision (ICCV), 2017.

[24] Marcella Cornia, Lorenzo Baraldi, Giuseppe Serra, and Rita Cucchiara, “A Deep Multi-Level Network for Saliency Prediction,” in International Conference on Pattern Recognition (ICPR), 2016.

[25] Nian Liu and Junwei Han, “A deep spatial contextual long-term recurrent convolutional network for saliency detection,” IEEE Transactions on Image Processing, vol. 27, no. 7, pp. 3264–3274, 2018.

[26] Marcella Cornia, Lorenzo Baraldi, Giuseppe Serra, and Rita Cucchiara, “Predicting Human Eye Fixations via an LSTM-based Saliency Attentive Model,” IEEE Trans- actions on Image Processing, vol. 27, no. 10, pp. 5142–

5154, 2018.

[27] Sen Jia, “EML-NET: an expandable multi-layer network for saliency prediction,” CoRR, vol.

abs/1805.01047, 2018.

[28] Wallis T. Kuemmerer M. and Bethge M., “Information- theoretic model comparison unifies saliency metrics,”

Proceedings of the National Academy of Science, vol.

112, no. 52, pp. 16054–16059, Oct 2015.

[29] Matthias K¨ummerer, Thomas SA Wallis, and Matthias Bethge, “Deepgaze ii: Reading fixations from deep features trained on object recognition,” arXiv preprint arXiv:1610.01563, 2016.

[30] Alex Kendall and Roberto Cipolla, “Geometric loss functions for camera pose regression with deep learn- ing,” in Proceedings of the IEEE Conference on Com- puter Vision and Pattern Recognition (CVPR), 2017.

[31] Justin Johnson, Alexandre Alahi, and Li Fei-Fei, “Per- ceptual losses for real-time style transfer and super- resolution,” in European Conference on Computer Vi- sion, 2016.

[32] T. Lin, P. Goyal, R. Girshick, K. He, and P. Dollr, “Fo- cal loss for dense object detection,” in 2017 IEEE Inter- national Conference on Computer Vision (ICCV), Oct 2017, pp. 2999–3007.

[33] Karen Simonyan and Andrew Zisserman, “Very deep convolutional networks for large-scale image recogni- tion,” arXiv preprint arXiv:1409.1556, 2014.

[34] Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam, “Rethinking atrous convolution for semantic image segmentation,” arXiv preprint arXiv:1706.05587, 2017.

[35] Tilke Judd, Krista Ehinger, Fr´edo Durand, and Antonio Torralba, “Learning to predict where humans look,”

in 12th international conference on Computer Vision.

IEEE, 2009, pp. 2106–2113.

[36] Junting Pan, Cristian Canton, Kevin McGuinness, Noel E O’Connor, Jordi Torres, Elisa Sayrol, and Xavier Giro-i Nieto, “Salgan: Visual saliency prediction with generative adversarial networks,” arXiv preprint arXiv:1701.01081, 2017.

[37] Wenguan Wang, Jianbing Shen, and Ling Shao, “Video salient object detection via fully convolutional net- works,” IEEE Transactions on Image Processing, vol.

27, no. 1, pp. 38–49, 2018.

[38] Olivier Le Meur and Thierry Baccino, “Methods for comparing scanpaths and saliency maps: strengths and weaknesses,” Behavior Research Method, vol. 45, no. 1, pp. 251–266, 2013.

[39] Robert J Peters, Asha Iyer, Laurent Itti, and Christof Koch, “Components of bottom-up gaze allocation in natural images,” Vision research, vol. 45, no. 18, pp.

2397–2416, 2005.

[40] Ecker A.S. Bethge M. Gatys, L.A., “A neural algorithm of artistic style,” in arXivpreprint, 2015.

[41] A. Garcia-Diaz, X. R. Fdez-Vidal, X. M. Pardo, and R. Dosil, “Saliency from hierarchical adaptation through decorrelation and variance normalization,” Im- age and Vision Computing, vol. 30, no. 1, pp. 51 – 64, 2012.

[42] Junting Pan, Elisa Sayrol, Xavier Giro-i Nieto, Kevin McGuinness, and Noel E O’Connor, “Shallow and deep convolutional networks for saliency prediction,” in Pro- ceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 598–606.

[43] Ali Borji and Laurent Itti, “Cat2000: A large scale fixation dataset for boosting saliency research,” arXiv preprint arXiv:1505.03581, 2015.

[44] Chengyao Shen and Qi Zhao, “Webpage saliency,” in ECCV. 2014, IEEE.

[45] Wang Y. Chang G.J., Zhang Y., “An element sensi-

tive saliency model with position prior learning for web

pages,” in ICIAI, 2018.

Références

Documents relatifs

Ensuite, nous avons comparé des outils tout-en-un, des technologies de base de données, des langages de programmation ainsi que des librairies graphiques pour trouver une

C’est ici ce qui m’intéresse particulièrement puisque je souhaite me rendre dans quatre classes d’école primaire afin de pouvoir y interroger les enseignants sur leur

We added the matrix rank and its properties to Thiemann and Yamada’s matrix library [41], generalized the definition of the Borel measure by Hölzl and Himmel- mann [24], and

It’s hard to write a poem when there’s no one person you’re thinking it’s for, it’s just a dirty day and a long walk looking for strangers that look famous so you feel like you

The main goal of the VAriability of vertical and tropHIc transfer of diazotroph derived N in the south wEst Pacific (VAHINE) project was to study the fate of nitrogen newly fixed

are described at different levels of approximation: (i) far from the drop boundary the flow is known analytically (inside and outside the drop) ; (ii) in the static meniscus (close

The criterion for solar radiative effects to occur was to heat the exterior surface until the surface temperature on the cladding (i.e. cementitious board) reached, e.g., 13 °C

Our test bench to assess scalability for DL-based steganalysis Choice of the network for JPEG steganalysis.. Choice of the database Choice of