• Aucun résultat trouvé

Steganalysis by Ensemble Classifiers with Boosting by Regression, and Post-Selection of Features

N/A
N/A
Protected

Academic year: 2022

Partager "Steganalysis by Ensemble Classifiers with Boosting by Regression, and Post-Selection of Features"

Copied!
26
0
0

Texte intégral

(1)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012

Steganalysis by Ensemble Classifiers with Boosting by Regression, and Post-Selection of Features

Marc Chaumont, Sarra Kouider LIRMM, Montpellier, France

October 2, 2012

IEEE International Conference on Image Processing 2012, Sept. 30 - Oct. 3 2012, Orlando, USA.

(2)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Preamble

Outline

1

Preamble

2

The Kodovsky’s Ensemble Classifiers

3

Boosting by regression

4

Post-selection of features

5

Experiments

6

Conclusion

(3)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Preamble

Steganography vs Steganalysis

(4)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Preamble

The proposition

An Improvement of a state-of-the-art steganalyzer

PE &

of the steganalyzer THANKS TO

boosting by regression of low complexity,

post-selection of features of low complexity.

(5)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 The Kodovsky’s Ensemble Classifiers

Outline

1

Preamble

2

The Kodovsky’s Ensemble Classifiers

3

Boosting by regression

4

Post-selection of features

5

Experiments

6

Conclusion

(6)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 The Kodovsky’s Ensemble Classifiers

Notable properties

Appeared during BOSS challenge (sept. 2010 - jan. 2011), Performances

to SVM,

Scalable regarding the dimension of the features vector, Low computational complexity,

Low memory complexity, Easily parallelizable.

J. Kodovsk´y, J. Fridrich, and V. Holub,

“Ensemble classifiers for steganalysis of digital media,”

IEEE Transactions on Information Forensics and Security, vol. 7, no. 2, pp. 432–

444, 2012.

(7)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 The Kodovsky’s Ensemble Classifiers

Definition of a weak classifier

Ensemble Classifiers is made of

L

weak classifiers Let

x∈Rd

a feature vector,

A weak classifier,

hl

, returns 0 for cover, 1 for stego :

hl

:

Rd → {0,

1}

x → hl

(x)

(8)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 The Kodovsky’s Ensemble Classifiers

How does classification work?

1

Take an image to analys (i.e. classify in cover or stego),

2

Extract the features vector

x∈Rd

,

3

Decide to classify cover or stego (majority vote):

C

(x) =

0 if

Pl=L

l=1hl

(x)

≤L/2,

1 otherwise.

(9)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Boosting by regression

Outline

1

Preamble

2

The Kodovsky’s Ensemble Classifiers

3

Boosting by regression

4

Post-selection of features

5

Experiments

6

Conclusion

(10)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Boosting by regression

Weighting the weak classifiers

1

Take an image to analys (i.e. classify in cover or stego),

2

Extract the features vector

x∈Rd

,

3

Decide to classify cover or stego (majority vote):

C

(x) =

0 if

Pl=L

l=1hl

(x)

≤L/2,

1 otherwise.

The classification (steganalysis) process was:

(11)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Boosting by regression

Weighting the weak classifiers

BUT : some weak classifiers are less efficient than others.

THEN : introduce weights !

1

Take an image to analys (i.e. classify in cover or stego),

2

Extract the features vector

x∈Rd

,

3

Decide to classify cover or stego (majority vote):

C

(x) =

0 if

Pl=L

l=1hl

(x)

≤L/2,

1 otherwise.

The classification (steganalysis) process was:

(12)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Boosting by regression

Weighting the weak classifiers

1

Take an image to analys (i.e. classify in cover or stego),

2

Extract the features vector

x∈Rd

,

3

Decide to classify cover or stego (weighted vote):

C

(x) =

(

0 if

Pl=L

l=1αlhl

(x)

Pl=L l=1αl

2 ,

1 otherwise.

The classification (steganalysis) process is now:

(13)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Boosting by regression

Weighting the weak classifiers

How to calculate those weights with a small computational complexity ?

1

Take an image to analys (i.e. classify in cover or stego),

2

Extract the features vector

x∈Rd

,

3

Decide to classify cover or stego (weighted vote):

C

(x) =

(

0 if

Pl=L

l=1αlhl

(x)

Pl=L l=1αl

2 ,

1 otherwise.

The classification (steganalysis) process is now:

(14)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Boosting by regression

Analytic expression of the weights

During learning step:

l}

= arg

l}

min

PE.

simplify

PE

expression, least squares problem

linear system

A.X

=

B

with

X

the weights :

Ai,j

=

n=N

X

n=1

hi

(x

n

)h

j

(x

n

),

Bi

=

n=N

X

n=1

hi

(x

n

)y

n.

... solved thanks to a library of linear algebra.

Order of complexity unchanged.

(15)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Post-selection of features

Outline

1

Preamble

2

The Kodovsky’s Ensemble Classifiers

3

Boosting by regression

4

Post-selection of features

5

Experiments

6

Conclusion

(16)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Post-selection of features

Reducing the dimension with few computations

Selection of features:

Pre-selection may cost a lot.

What about post-selection?

1

Take an image to analys (i.e. classify in cover or stego),

2

Extract the features vector

x∈Rd

,

3

Decide to classify cover or stego (weighted vote):

C

(x) =

(

0 if

Pl=L

l=1αlhl

(x)

Pl=L l=1αl

2 ,

1 otherwise.

Remember:

The classification (steganalysis) process is now:

(17)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Post-selection of features

Order of complexity unchanged.

Algorithm :

1

Compute a

score

for each feature;

first database reading,

2

Define an order of selection of the features,

3

Find the best subset (lowest

PE

);

second database reading.

Once a weak classifier learned : suppress the features reducing

PE

:

(18)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Experiments

Outline

1

Preamble

2

The Kodovsky’s Ensemble Classifiers

3

Boosting by regression

4

Post-selection of features

5

Experiments

6

Conclusion

(19)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Experiments

Experimental conditions

10 000 greyscale images (512×512, BOSS database), The same 10 000 embedded at 0.4 bpp with HUGO,

Feature vector dimension

d

= 5330 features (HOLMES subset), 5 different splits, 5 different seeds,

HUGO: “Using High-Dimensional Image Models to Perform Highly Undetectable Steganography”

T. Pevn´y, T. Filler, and P. Bas, in Information Hiding, IH’2010.

HOLMES: “Steganalysis of Content-Adaptive Steganography in Spatial Domain”

J. Fridrich, J. Kodovsk´y, V. Holub, and M. Goljan, in Information Hiding, IH’2011.

(20)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Experiments

Steganalysis results

(21)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Experiments

Steganalysis results

Recall increase = 1.7%

Same computational complexity order

(22)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Conclusion

Outline

1

Preamble

2

The Kodovsky’s Ensemble Classifiers

3

Boosting by regression

4

Post-selection of features

5

Experiments

6

Conclusion

(23)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Conclusion

Summary

Two propositions for the Kodovsk´ y steganalyzer:

boosting by regression, post-selection of features.

Significant recall increase (1.7%)

No change in computational complexity order

(24)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Conclusion

Annex: Metrics (1)

Distance between the two classes:

c1(l)[j] = 1[j]µ0[j]|

q

σ21[j] +σ02[j] .

Influence of a feature on the final correlation/decision ( = dot product) used to classify:

c2(l)[j] =

i=N

X

i=1

count(x(l)i [j],w(l)[j],yi),

with:

count(x,w,y) =

1 if [(x.w>0 andy= 1) or (x.w<0 andy= 0)], 0 otherwise.

c3(l)[j] =

i=N

X

i=1

count(x(l)i [j],w(l)[j],yi) Pk=dred

k=1 count(x(l)i [k],w(l)[k],yi) .

(25)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Conclusion

Annex: Metrics (2)

Feature correlation with the class:

c4(l)[j] = corr(x(l)[j],y)

=

Pi=N i=1

x(l)i [j]x(l)[j]

(yiy) r

Pi=N i=1

x(l)i [j]x(l)[j]

2q Pi=N

i=1 (yiy)2

.

Feature correlation with the weak classifier:

c5(l)[j] =corr(x(l)[j].w(l)[j],y).

(26)

Steganalysis by Ensemble Classifiers - Marc Chaumont - ICIP’2012 Conclusion

Annex: P

E

in the Boosting by Regression

During learning step:

l}= arg

l}

minPE.

PE = 1 N

i=N

X

i=1

f

l=L

X

l=1

αlhl(xi)

!

−yi

! .

withf a thresholding function defined by:

f :R → {0,1}

x → f(x) = (

0 ifx≤ Pl=Ll=12αl, 1 otherwise.

Let’s simplify,PE:

PE ≈ 1 N

i=N

X

i=1 l=L

X

l=1

αlhl(xi)−yi

!2 .

⇒least squares problem ... solved thanks to a library of linear algebra.

Références

Documents relatifs

We introduce in Section 2 a general framework for studying the algorithms from the point of view of functional optimization in an L 2 space, and prove in Section 3 their convergence

Given an input image, the classifier of which learning has been achieved on images with ”similar” feature values will be less sensitive to the cover-source mismatch

Figure 2 presents the results we obtained (with a 5 fold cross-validation procedure [17]) over 50 iterations with (i) 2-B OOST , (ii) the two single boosted weak learners, and

In this paper, we take advantages of two machine learning approaches, gradient boosting and random Fourier features, to derive a novel algorithm that jointly learns a

In this respect, the objective of this work is to estimate the degree of confidence in the identification of nuclear transients by a diagnostic system based on a

Of these, Kohler ’s rule is by far the fastest ranking heuristic, yet also the less satisfactory in terms of ordinal correlation with a corresponding pairwise outranking relation..

Motivation Ranking-by-choosing q-tiles sorting Boosting Kohler’s rule Conclusion.. On boosting Kohler’s ranking-by-choosing rule with a

To relate instability with performance of bagging, as Breiman does in [11] to show that bagging improves only unstable methods, we compute, for each data set, the instability index