• Aucun résultat trouvé

Computing Dependencies between DCT Coefficients for Natural Steganography in JPEG Domain

N/A
N/A
Protected

Academic year: 2021

Partager "Computing Dependencies between DCT Coefficients for Natural Steganography in JPEG Domain"

Copied!
7
0
0

Texte intégral

(1)

HAL Id: hal-02165866

https://hal.archives-ouvertes.fr/hal-02165866v2

Submitted on 13 Sep 2019

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

Computing Dependencies between DCT Coefficients for

Natural Steganography in JPEG Domain

Théo Taburet, Patrick Bas, Jessica Fridrich, Wadih Sawaya

To cite this version:

Théo Taburet, Patrick Bas, Jessica Fridrich, Wadih Sawaya.

Computing Dependencies between

DCT Coefficients for Natural Steganography in JPEG Domain. IH-MMSec, Jul 2019, Paris, France.

�10.1145/3335203.3335715�. �hal-02165866v2�

(2)

for Natural Steganography in JPEG Domain

Théo Taburet

Univ. Lille, CNRS, Centrale Lille, UMR 9189, CRIStAL Lille, France

theo.taburet@centralelille.fr

Patrick Bas

Univ. Lille, CNRS, Centrale Lille, UMR 9189, CRIStAL Lille, France patrick.bas@centralelille.fr

Jessica Fridrich

Department of ECE Binghamton fridrich@binghamton.edu

Wadih Sawaya

IMT Lille-Douais, Univ. Lille, CNRS, CRIStAL Lille, France

wadih.sawaya@imt-lille-douai.fr

ABSTRACT

This short paper is an extension of a family of embedding schemes called Natural Steganography, which embeds a message by mim-icking heteroscedastic sensor noise in the JPEG domain. Under the assumption that the development from RAW uses linear de-mosaicking, we derive a closed-form for the covariance matrix of DCT coefficients from 3 × 3 JPEG blocks. This computation relies on a matrix formulation of all steps involved in the development pipeline, which includes demosaicking, conversion to luminance, DCT transform, and reordering. This matrix is then used for pseudo-embedding in the JPEG domain on four lattices of 8 × 8 DCT blocks. The results obtained with the computed covariance matrix are con-trasted with the results previously obtained with the covariance matrix estimated using Monte Carlo sampling and scaling. The empirical security using DCTR features at JPEG quality 100 in-creased from PE = 14% using covariance estimation and scaling to PE= 43% using the newly derived analytic form.

CCS CONCEPTS

• Security and privacy → Domain-specific security and privacy architectures; Intrusion/anomaly detection and malware mit-igation; Malware and its mitmit-igation;

KEYWORDS

Digital image steganography, JPEG domain, sensor noise, image processing pipeline, covariance

ACM Reference Format:

Théo Taburet, Patrick Bas, Jessica Fridrich, and Wadih Sawaya. 2019. Com-puting Dependencies between DCT Coefficients for Natural Steganography in JPEG Domain. In Proceedings of ACM Conference (Conference’17). ACM, New York, NY, USA, 6 pages. https://doi.org/10.1145/nnnnnnn.nnnnnnn

1

INTRODUCTION

High capacity steganography, i. e., steganographic schemes with high empirical security at high embedding rates, can be designed us-ing various strategies, such as model-based embeddus-ing, generative steganography, and side-informed embedding.

Conference’17, July 2017, Washington, DC, USA © 2019 Association for Computing Machinery. ACM ISBN 978-x-xxxx-xxxx-x/YY/MM...$15.00 https://doi.org/10.1145/nnnnnnn.nnnnnnn

Model-based embedding first estimates a model of the cover and then embeds the message by either preserving the model [14] or by minimizing the statistical impact of embedding [15]. To some extent, the class of generative steganography [8] relies on the same rationale by sampling the content directly to carry a specific mes-sage. In [19], stego text is generated using LSTM and in [18] stego images are generated using auto-regressive models. One advantage of generative steganography is the fact that the steganographic capacity can be close to the entropy H(X) of the source X by em-ploying codes, such as the Syndrome-Trellis Codes (STCs) [7], which operate close to the corresponding rate-distortion bound.

Another possibility to increase the empirical security of an em-bedding scheme is to use side-information in the form of a pre-cover, a high-resolution version of the cover. Such methods use the quanti-zation error computed during the conversion from the pre-cover to cover, large quantization errors inducing high embedding change probabilities. Notable examples of side-informed schemes include Perturbed Quantization [9], SI-UNIWARD [11], and other adaptions of cost-based embedding schemes [5]. Note that if the pre-cover is not available, one can try estimating it [17] or use multiple shots of the same image [6] to estimate “fluctuating” coefficients, i. e., coefficients associated with large quantization errors or with sensor noise of large variance.

In this paper, we propose another high capacity steganographic scheme for the JPEG domain that relies on the principles of model-based and side-informed steganography. It also belongs to the class of Natural Steganography (NS) [1, 2, 4, 16] that generates the stego signal by mimicking the sensor noise associated with a larger ISO sensitivity. Section 2 reviews the principle of NS and its recent implementation in the JPEG domain. Section 3 details the main contribution of this paper, the analytic computation of the covari-ance matrix among neighboring DCT coefficients after a linear development pipeline from RAW to JPEG. Section 4 compares the performances of the proposed scheme to previous implementations.

2

NATURAL STEGANOGRAPHY

Throughout this paper, we use capital letters for random variables and the corresponding lower-case symbols for their realizations. Matrices are typed in upper-case and vectors in lower-case boldface font. Matrix transpose will be denoted with a superscript “t”.

(3)

Conference’17, July 2017, Washington, DC, USA Taburet, Bas, Fridrich & Sawaya

2.1

The principle of NS

The main idea behind NS relies on the principle of cover-source switching [1], which consists of generating stego content that is statistically similar to the cover source acquired at a different cam-era setting. Here, the source is defined by the shot noise at the photo-site level due to the photon counting process occurring on CCD or CMOS sensors. This noise is independent across photosites and only depends on the sensor model and the ISO setting.

After embedding, the stego image generated from a cover ac-quired at sensitivity ISO1should have the same statistical properties as a cover image acquired at sensitivity ISO2in order to guarantee high empirical security. The shot noise N(i)at ISOi, i ∈ {1, 2}, fol-lows a zero-mean Gaussian distribution with variance determined by the luminance of the noiseless photo-site value µ:

N(i)∼ N (0, aiµ + bi), (1) where the pair of parameters (ai,bi)depends only on the sensor and ISOi. Consequently, one can show [1] that, at the photo-site level, the stego image mimicking an image acquired at sensitivity ISO2is distributed as:

S ∼ N(0, (a2−a1)µ + b2−b1). (2) This idea was the cornerstone of NS embedding schemes proposed for monochrome sensors in the spatial domain assuming a sim-plified development pipeline that includes quantization, gamma transform, and downscaling [1, 2]. The method was shown to pro-vide high steganographic capacity with high empirical security.

2.2

Embedding algorithm in the JPEG domain

In the current paper, we study NS when the pre-cover is acquired with a color sensor, developed with linear demosaicking, converted to grayscale, and JPEG compressed. The embedding is executed directly in the JPEG domain by manipulating the quantized DCT coefficients.

The difference between monochrome and color sensors was studied in [4] with the conclusion that independent embedding on each DCT coefficient offers high empirical security for mono-chrome sensors, but not for color sensors. This is due to the fact that demosaicking introduces dependencies among neighboring DCT coefficients. When these dependencies are not taken into ac-count, the embedding scheme becomes highly detectable at high JPEG Quality Factors (QF). To overcome this problem, the authors of paper [16] modeled these dependencies using the multi-variate Gaussian model with the covariance matrix of the stego signal in the DCT domain ˆΣ estimated from a constant-luminosity RAW image altered by shot-noise. The embedding was then designed to respect the required covariance among stego DCT coefficients. The empirical security of a simulated-embedding scheme1was indeed larger than when the embedding was assuming independent DCT coefficients [4].

The embedding scheme is summarized in Figure 1 and can be decomposed into different steps described below:

1As commonly done for cost-based embedding schemes, here the message is not

embedded but the embedding changes are simulated by computing the embedding change probabilities and sampling according to them.

DCT Coeff. Covariance matrix Scaling MGD Sampling on four lattices Pre-cover (Raw) ISO1 <latexit sha1_base64="FWnwYj7X8eZd3FxoVOpVs+NmY9U=">AAAB7HicbVBNS8NAEJ34WetX1aOXxSJ4KkkV9FjwoicrmrbQhrLZTtqlm03Y3Qil9Dd48aCIV3+QN/+N2zYHbX0w8Hhvhpl5YSq4Nq777aysrq1vbBa2its7u3v7pYPDhk4yxdBniUhUK6QaBZfoG24EtlKFNA4FNsPh9dRvPqHSPJGPZpRiENO+5BFn1FjJv32463rdUtmtuDOQZeLlpAw56t3SV6eXsCxGaZigWrc9NzXBmCrDmcBJsZNpTCkb0j62LZU0Rh2MZ8dOyKlVeiRKlC1pyEz9PTGmsdajOLSdMTUDvehNxf+8dmaiq2DMZZoZlGy+KMoEMQmZfk56XCEzYmQJZYrbWwkbUEWZsfkUbQje4svLpFGteOeV6v1FuebmcRTgGE7gDDy4hBrcQB18YMDhGV7hzZHOi/PufMxbV5x85gj+wPn8AQGMjhk=</latexit> ISO1 <latexit sha1_base64="FWnwYj7X8eZd3FxoVOpVs+NmY9U=">AAAB7HicbVBNS8NAEJ34WetX1aOXxSJ4KkkV9FjwoicrmrbQhrLZTtqlm03Y3Qil9Dd48aCIV3+QN/+N2zYHbX0w8Hhvhpl5YSq4Nq777aysrq1vbBa2its7u3v7pYPDhk4yxdBniUhUK6QaBZfoG24EtlKFNA4FNsPh9dRvPqHSPJGPZpRiENO+5BFn1FjJv32463rdUtmtuDOQZeLlpAw56t3SV6eXsCxGaZigWrc9NzXBmCrDmcBJsZNpTCkb0j62LZU0Rh2MZ8dOyKlVeiRKlC1pyEz9PTGmsdajOLSdMTUDvehNxf+8dmaiq2DMZZoZlGy+KMoEMQmZfk56XCEzYmQJZYrbWwkbUEWZsfkUbQje4svLpFGteOeV6v1FuebmcRTgGE7gDDy4hBrcQB18YMDhGV7hzZHOi/PufMxbV5x85gj+wPn8AQGMjhk=</latexit> Stego (JPEG) DCT Coeff. Quantization Demosaicking / Luminance / DCT ˆ ⌃

<latexit sha1_base64="(null)">(null)</latexit><latexit sha1_base64="(null)">(null)</latexit><latexit sha1_base64="(null)">(null)</latexit><latexit sha1_base64="(null)">(null)</latexit>

Figure 1: Overview of the embedding scheme presented in [16].

(1) Estimation of the covariance matrix ˆΣ between DCT coefficients of3 × 3 neighboring 8 × 8 blocks. Since it is computationally infeasible to estimate the covariance matrix for each block of the cover image, the estimation is performed on a constant-luminosity RAW image with photo-site values µ = 212coded on 14 bits and corrupted with the stego signal S given by (2). This estimation uses No= 6×104observations of 24×24 DCT coefficients obtained from a developed (3480× 4640) RAW image.

(2) Beginning of the development (demosaicking, lumi-nance transform, and DCT).The cover RAW image fol-lows a classical development pipeline to generate grayscale JPEG images. After demosaicking, the standard RGB to lu-minance transform2given by: yl = 0.299 yr + 0.587 yд+ 0.114 yb, is applied, followed by a 2D-DCT transform on 8 × 8 blocks.

(3) Scaling of ˆΣ. Since we assume that the development is linear, and in order to take into account the conversion from RGB to luminance, the covariance matrix associated with each block is scaled as ˆΣ′= γ ˆΣ, where γ represents the scaling factor given by:

γ = 0.2992(a¯xr+ b) + 0.5872(a¯xд+ b) + 0.1142(a¯xb+ b) (0.2992+ 0.5872+ 0.1142)(a212+ b) , (3) where ¯xr, ¯xд, and ¯xb represent, respectively, the average photo-site value of the red, green, and blue component of the block that is sampled.

(4) Sampling on four lattices. In [16], the authors have shown that for this development, the stego signal generated on two non 8-connected blocks is independent and that the dependencies between 8-connected blocks are solely due to demosaicking. Consequently, we can use four lattices {Λ1, Λ2, Λ3, Λ4}depicted in Figure 2 to sample the stego signal in the DCT domain. DCT blocks belonging to lattice Λ1are sampled independently, blocks belonging to lattice Λ2are sampled conditionally on the four diagonal blocks, blocks belonging to lattice Λ3are sampled conditionally on the four vertical and horizontal blocks, and blocks of Λ4are sampled conditionally on the 8 surrounding blocks. We can show that for a linear development, the distribution of the stego signal for each lattice Λj (j ∈ {1, .., 4}) and block i follows a Multivariate Gaussian Distribution (MGD):

N (mi, j, Σi, j), (4) where for j ∈ {2, 3, 4}, the expectation vector mi, jand covari-ance matrix Σi, jof the conditional distribution are computed

2Without loss of generality we assume that no other color transform is applied, however

(4)

for Natural Steganography in JPEG Domain Conference’17, July 2017, Washington, DC, USA using the Schur complement of the estimated covariance

ma-trix ˆΣ (see [13]).

(5) JPEG quantization. The simulated stego signal is quan-tized using the JPEG quantization matrix for a given Quality Factor (QF). One can also compute the pmf of the stego signal on JPEG coefficients at the expense of increased complexity. Note that in order to perform practical embedding, one must use 64 sub-lattices in each block (one for each DCT coeffi-cient), this can be done by computing the conditional pmf and the associated costs (see [16]). The pmf πq,i for each coefficient i considering a Q-arry alphabet and symbol q is also used to estimate the average payload embedded in each coefficient by computing the entropy

H(πq,i)= − Q Õ q=1

πq,ilog2πq,i. (5) As detailled in [16], the πq,i are computed by dividing the nor-mal marginal distribution for coefficient i into Q bins and then computing the pmf. 1 3 1 3 1 4 1 2 1 2 1 3 3 4 4 4 1 2 3 : ⇤1 : ⇤2 : ⇤3 4 : ⇤4

<latexit sha1_base64="(null)">(null)</latexit><latexit sha1_base64="(null)">(null)</latexit><latexit sha1_base64="(null)">(null)</latexit><latexit sha1_base64="(null)">(null)</latexit>

1 2 1 2 1 3 3 4 4 8 8

Figure 2: The four latticesΛ1,Λ2,Λ3andΛ4, used for embed-ding.

The proposed simulated embedding scheme has pros and cons. On the one hand, it offers good empirical security for medium JPEG QFs (85 and 75) and for linear or close to linear demosaicking algorithms (bilinear or VNG, see Table 2 in [16]). On the other hand, the estimation of the covariance matrix combined with its scaling (3) are only approximations that decrease the empirical security of the whole scheme, especially for high QFs.

We consequently derive in this paper a closed-form of the co-variance matrix Σ for bilinear demosaicking. This matrix is directly computed from the photo-site values of the RAW cover image and does not need to be scaled. This approach is validated in section 4 by showing the results are equivalent to the ones obtained by simulated embedding at the photo-site level.

3

ANALYTIC DERIVATION OF THE

COVARIANCE MATRIX

A RAW image captured using a color sensor can be developed using a variety of processes, which include the bilinear interpolation or non-linear gradient based interpolations, such as DCB or AAHD. Below, we consider the simplest case of bilinear demosaicking.

In order to compute the covariance matrix Σ of the stego sig-nal in the DCT domain, we need to decompose the development pipeline into a sequence of linear steps illustrated in Figure 3 and summarized below:

(1) Demosaicking, here we consider the Bayer CFA (Color Filter Array) pattern shown in Figure 4.

(2) Luminance averaging, we consider this process because the embedding is performed in the luminance domain. (3) Crops to 8k × 8k blocks, with k ∈ {1, 3}, this pre-processing

is needed before the DCT transform. The parameter k is used to compute the covariance matrices used either for lattice Λ1 (82×82covariance matrix), lattices Λ2and Λ3(5·(8)2×5·(8)2 covariance matrices), and Λ4((3 × 8)2× (3 × 8)2covariance matrix).

(4) Conversion from row scan to block scan, this permutation is also needed before the DCT transform.

(5) DCT transform.

These linear operations can be written as a sequence of matrix multiplications. From a flattened vector of observations obtained from k×k blocks of photo-sites, we can write this operation pipeline as a single matrix M, such as tI= M · y, where tIis the flattened DCT block corresponding to the observation vector y∈N(8k+2)2. This vector y is an unfolded (8k + 2) × (8k + 2) matrix whose elements are photo-sites observations.

Vector : y2 N(k⇥8+2)2 <latexit sha1_base64="0A8sIhW2pkKckxlncJKktSGFkGc=">AAACKHicbVDLSsNAFJ3UV62vqks3wSJUhJJUwdKNBTeupIJ9QNOWyXTSDp1MwsxECEM+x42/4kZEkW79EidpF9p6YODMOfdy7z1uSImQljUzcmvrG5tb+e3Czu7e/kHx8Kgtgogj3EIBDXjXhQJTwnBLEklxN+QY+i7FHXd6m/qdJ8wFCdijjEPc9+GYEY8gKLU0LN44PpQT7qs2RjLgiVOvO/VMcz0VJw5h84+r7pOBKk8dSXwsahfV84GqJklhWCxZFSuDuUrsBSmBBZrD4rszClDkYyYRhUL0bCuUfQW5JIjipOBEAocQTeEY9zRlUI/rq+zQxDzTysj0Aq4fk2am/u5Q0Bci9l1dmW4tlr1U/M/rRdKr9RVhYSQxQ/NBXkRNGZhpauaIcJ0PjTWBiBO9q4kmkEMkdbZpCPbyyaukXa3YlxXr4arUsBZx5MEJOAVlYINr0AB3oAlaAIFn8Ao+wKfxYrwZX8ZsXpozFj3H4A+M7x/zdacA</latexit> Vector : y2 N(k⇥8+2)2 <latexit sha1_base64="0A8sIhW2pkKckxlncJKktSGFkGc=">AAACKHicbVDLSsNAFJ3UV62vqks3wSJUhJJUwdKNBTeupIJ9QNOWyXTSDp1MwsxECEM+x42/4kZEkW79EidpF9p6YODMOfdy7z1uSImQljUzcmvrG5tb+e3Czu7e/kHx8Kgtgogj3EIBDXjXhQJTwnBLEklxN+QY+i7FHXd6m/qdJ8wFCdijjEPc9+GYEY8gKLU0LN44PpQT7qs2RjLgiVOvO/VMcz0VJw5h84+r7pOBKk8dSXwsahfV84GqJklhWCxZFSuDuUrsBSmBBZrD4rszClDkYyYRhUL0bCuUfQW5JIjipOBEAocQTeEY9zRlUI/rq+zQxDzTysj0Aq4fk2am/u5Q0Bci9l1dmW4tlr1U/M/rRdKr9RVhYSQxQ/NBXkRNGZhpauaIcJ0PjTWBiBO9q4kmkEMkdbZpCPbyyaukXa3YlxXr4arUsBZx5MEJOAVlYINr0AB3oAlaAIFn8Ao+wKfxYrwZX8ZsXpozFj3H4A+M7x/zdacA</latexit> Raw samples from photosites

Interpolated color channels: (d(d<latexit sha1_base64="L95tRUJ0AI9gtzQO3teVs0dkFe8=">AAACMXicbVBNS8NAEN3Ur1q/qh69LBahRSlJK9hj0YM9SRX7AU0tm+2mXbrZhN2NUEL+khf/iXjpQRGv/gk3bUFbHRh4780MM/OcgFGpTHNipFZW19Y30puZre2d3b3s/kFT+qHApIF95ou2gyRhlJOGooqRdiAI8hxGWs7oKqm3HomQ1Of3ahyQrocGnLoUI6WlXraWtz2kho4b9Xt38Rn8YdcL7DIu2JTPuBPdxA9RvjI6LRVsRT0iy3Gml82ZRXMa8C+w5iAH5lHvZV/svo9Dj3CFGZKyY5mB6kZIKIoZiTN2KEmA8AgNSEdDjvSebjT9OIYnWulD1xc6uYJT9fdEhDwpx56jO5OT5XItEf+rdULlVroR5UGoCMezRW7IoPJhYh/sU0GwYmMNEBZU3wrxEAmElTY5McFafvkvaJaKVrlo3Z7nqubcjjQ4AscgDyxwAaqgBuqgATB4Aq/gDbwbz8bE+DA+Z60pYz5zCBbC+PoGFtSpcA==</latexit> RR, d, dGG, d, dBB))2 N2 N(8k+2)⇥3 (8k+2)⇥3 <latexit sha1_base64="L95tRUJ0AI9gtzQO3teVs0dkFe8=">AAACMXicbVBNS8NAEN3Ur1q/qh69LBahRSlJK9hj0YM9SRX7AU0tm+2mXbrZhN2NUEL+khf/iXjpQRGv/gk3bUFbHRh4780MM/OcgFGpTHNipFZW19Y30puZre2d3b3s/kFT+qHApIF95ou2gyRhlJOGooqRdiAI8hxGWs7oKqm3HomQ1Of3ahyQrocGnLoUI6WlXraWtz2kho4b9Xt38Rn8YdcL7DIu2JTPuBPdxA9RvjI6LRVsRT0iy3Gml82ZRXMa8C+w5iAH5lHvZV/svo9Dj3CFGZKyY5mB6kZIKIoZiTN2KEmA8AgNSEdDjvSebjT9OIYnWulD1xc6uYJT9fdEhDwpx56jO5OT5XItEf+rdULlVroR5UGoCMezRW7IoPJhYh/sU0GwYmMNEBZU3wrxEAmElTY5McFafvkvaJaKVrlo3Z7nqubcjjQ4AscgDyxwAaqgBuqgATB4Aq/gDbwbz8bE+DA+Z60pYz5zCBbC+PoGFtSpcA==</latexit> R, G, B <latexit sha1_base64="/qclm4hnjgt5/AYfwjIo5BuH2Wc=">AAAB7nicbVA9SwNBEJ2LX0n8ilraLAbBIoQ7LbQMWmgZxXxAcoS9vb1kyd7esbsnhCONjbWNhSK2Nv4ZO3+Nbi4pNPHBwOO9GWbmeTFnStv2l5VbWl5ZXcsXiusbm1vbpZ3dpooSSWiDRDySbQ8rypmgDc00p+1YUhx6nLa84cXEb91RqVgkbvUopm6I+4IFjGBtpNZNBV1W0HmvVLardga0SJwZKdcKD/7H932l3it9dv2IJCEVmnCsVMexY+2mWGpGOB0Xu4miMSZD3KcdQwUOqXLT7NwxOjSKj4JImhIaZerviRSHSo1Cz3SGWA/UvDcR//M6iQ7O3JSJONFUkOmiIOFIR2jyO/KZpETzkSGYSGZuRWSAJSbaJFQ0ITjzLy+S5nHVOana1yYNG6bIwz4cwBE4cAo1uII6NIDAEB7hGV6s2HqyXq23aWvOms3swR9Y7z8F/pGt</latexit> R, G, B <latexit sha1_base64="/qclm4hnjgt5/AYfwjIo5BuH2Wc=">AAAB7nicbVA9SwNBEJ2LX0n8ilraLAbBIoQ7LbQMWmgZxXxAcoS9vb1kyd7esbsnhCONjbWNhSK2Nv4ZO3+Nbi4pNPHBwOO9GWbmeTFnStv2l5VbWl5ZXcsXiusbm1vbpZ3dpooSSWiDRDySbQ8rypmgDc00p+1YUhx6nLa84cXEb91RqVgkbvUopm6I+4IFjGBtpNZNBV1W0HmvVLardga0SJwZKdcKD/7H932l3it9dv2IJCEVmnCsVMexY+2mWGpGOB0Xu4miMSZD3KcdQwUOqXLT7NwxOjSKj4JImhIaZerviRSHSo1Cz3SGWA/UvDcR//M6iQ7O3JSJONFUkOmiIOFIR2jyO/KZpETzkSGYSGZuRWSAJSbaJFQ0ITjzLy+S5nHVOana1yYNG6bIwz4cwBE4cAo1uII6NIDAEB7hGV6s2HqyXq23aWvOms3swR9Y7z8F/pGt</latexit> Luminance averaging DD <latexit sha1_base64="O0pZUMaUGwVPpKdPbKezmUZT8nA=">AAAB9XicbVDLSgMxFL3js9ZX1aWbYBFclRkVdFmwC5cV7APasWTSTBuaZIYko5Rh/sONC0Xc+i/u/Bsz7Sy09UDgcM693JMTxJxp47rfzsrq2vrGZmmrvL2zu7dfOThs6yhRhLZIxCPVDbCmnEnaMsxw2o0VxSLgtBNMbnK/80iVZpG8N9OY+gKPJAsZwcZKD32BzTgI08YgbWTZoFJ1a+4MaJl4BalCgeag8tUfRiQRVBrCsdY9z42Nn2JlGOE0K/cTTWNMJnhEe5ZKLKj201nqDJ1aZYjCSNknDZqpvzdSLLSeisBO5in1opeL/3m9xITXfspknBgqyfxQmHBkIpRXgIZMUWL41BJMFLNZERljhYmxRZVtCd7il5dJ+7zmXdTcu8tq3S3qKMExnMAZeHAFdbiFJrSAgIJneIU358l5cd6dj/noilPsHMEfOJ8/vpiSmw==</latexit> DD <latexit sha1_base64="O0pZUMaUGwVPpKdPbKezmUZT8nA=">AAAB9XicbVDLSgMxFL3js9ZX1aWbYBFclRkVdFmwC5cV7APasWTSTBuaZIYko5Rh/sONC0Xc+i/u/Bsz7Sy09UDgcM693JMTxJxp47rfzsrq2vrGZmmrvL2zu7dfOThs6yhRhLZIxCPVDbCmnEnaMsxw2o0VxSLgtBNMbnK/80iVZpG8N9OY+gKPJAsZwcZKD32BzTgI08YgbWTZoFJ1a+4MaJl4BalCgeag8tUfRiQRVBrCsdY9z42Nn2JlGOE0K/cTTWNMJnhEe5ZKLKj201nqDJ1aZYjCSNknDZqpvzdSLLSeisBO5in1opeL/3m9xITXfspknBgqyfxQmHBkIpRXgIZMUWL41BJMFLNZERljhYmxRZVtCd7il5dJ+7zmXdTcu8tq3S3qKMExnMAZeHAFdbiFJrSAgIJneIU358l5cd6dj/noilPsHMEfOJ8/vpiSmw==</latexit> Pixel selection

Permutation for bloc scan

S<latexit sha1_base64="7rq2AfEqKy9hYn8lZB4X/b3juaU=">AAAB8XicbVC7SgNBFL3rM8ZXfHQ2g0GwCrNaaGfAQsuI5oHJEmYns8mQ2dllZlaIS/7CxkIRW8HKL7Gz9E+cTVJo4oELh3Pu5Z57/VhwbTD+cubmFxaXlnMr+dW19Y3NwtZ2TUeJoqxKIxGphk80E1yyquFGsEasGAl9wep+/zzz63dMaR7JGzOImReSruQBp8RY6bYVEtPzg/R62C4UcQmPgGaJOyHFs4/774v33bTSLny2OhFNQiYNFUTrpotj46VEGU4FG+ZbiWYxoX3SZU1LJQmZ9tJR4iE6sEoHBZGyJQ0aqb8nUhJqPQh925kl1NNeJv7nNRMTnHopl3FimKTjRUEikIlQdj7qcMWoEQNLCFXcZkW0RxShxj4pb5/gTp88S2pHJfe4hK9wsYxhjBzswT4cggsnUIZLqEAVKEh4gCd4drTz6Lw4r+PWOWcyswN/4Lz9AOQhlLU=</latexit> P <latexit sha1_base64="pxkiGLtCEsDe9kXt4cnCYtzEvwc=">AAAB8XicbVC7SgNBFL0bXzG+4qOzGQyCVZjVQjsDFlpGMA9MljA7mU2GzM4uM7NCXPIXNhaK2ApWfomdpX/ibJJCEw9cOJxzL/fc68eCa4Pxl5NbWFxaXsmvFtbWNza3its7dR0lirIajUSkmj7RTHDJaoYbwZqxYiT0BWv4g4vMb9wxpXkkb8wwZl5IepIHnBJjpdt2SEzfD9LqqFMs4TIeA80Td0pK5x/335fve2m1U/xsdyOahEwaKojWLRfHxkuJMpwKNiq0E81iQgekx1qWShIy7aXjxCN0aJUuCiJlSxo0Vn9PpCTUehj6tjNLqGe9TPzPayUmOPNSLuPEMEkni4JEIBOh7HzU5YpRI4aWEKq4zYponyhCjX1SwT7BnT15ntSPy+5JGV/jUgXDBHnYhwM4AhdOoQJXUIUaUJDwAE/w7Gjn0XlxXietOWc6swt/4Lz9AN+SlLI=</latexit> DCT transform l = DD·y <latexit sha1_base64="5FMsYwtzmWkm40YReeLCXBUjmkk=">AAACEnicbZDLSsNAFIYnXmu9RV26GSyCbkqigm6Egl24rGAv0IQwmU7aoZOZMDMRQsgzuPFV3LhQxK0rd76NSZuCtv4w8PGfc5hzfj9iVGnL+jaWlldW19YrG9XNre2dXXNvv6NELDFpY8GE7PlIEUY5aWuqGelFkqDQZ6Trj2+KeveBSEUFv9dJRNwQDTkNKEY6tzzz1AmRHvlByrLrGTa9tJk5eCB0NrOSrOqZNatuTQQXwS6hBkq1PPPLGQgch4RrzJBSfduKtJsiqSlmJKs6sSIRwmM0JP0cOQqJctPJSRk8zp0BDITMH9dw4v6eSFGoVBL6eWexopqvFeZ/tX6sgys3pTyKNeF4+lEQM6gFLPKBAyoJ1izJAWFJ810hHiGJsM5TLEKw509ehM5Z3T6vW3cXtYZVxlEBh+AInAAbXIIGuAUt0AYYPIJn8ArejCfjxXg3PqatS0Y5cwD+yPj8AZICnpg=</latexit> ⌃ =E[tI· ttI] <latexit sha1_base64="l5o60pJUNzsDSeO3RtdhJ4l4Uqw=">AAACL3icbVDLSsNAFJ34tr6iLt0Ei+CqJCroRhBE0V1FawtJLJPJpA6dZMLMjVCG/JEbf6UbEUXc+hdO2i58HRjmcM693HtPlHOmwHVfrKnpmdm5+YXF2tLyyuqavb5xq0QhCW0RwYXsRFhRzjLaAgacdnJJcRpx2o76p5XffqBSMZHdwCCnYYp7GUsYwWCkrn0eRILHapCaTwfXrJfi8jhIMdxHkT4r/TFNNHT1ZVkGJBbwQ7rTUJZhrWvX3YY7gvOXeBNSRxM0u/YwiAUpUpoB4Vgp33NzCDWWwAinZS0oFM0x6eMe9Q3NcEpVqEf3ls6OUWInEdK8DJyR+r1D41RVJ5nKalf126vE/zy/gOQo1CzLC6AZGQ9KCu6AcKrwnJhJSoAPDMFEMrOrQ+6xxARMxFUI3u+T/5LbvYa333CvDuon7iSOBbSFttEu8tAhOkEXqIlaiKBHNESv6M16sp6td+tjXDplTXo20Q9Yn1+K66u5</latexit> tI= (DT· P · S · DD)· y <latexit sha1_base64="4Wh46FO1Q7xBrlZNJ5HEzgyje2E=">AAACSXicbVC7SgNBFJ1NfMT4ilraLAYhNmFXBW2EgCm0i5gXJGGdnczGIbMPZu4KYZnfs7Gz8x9sLBSxcjYPMIkXBs459x7uneNGnEmwrDcjk11ZXVvPbeQ3t7Z3dgt7+00ZxoLQBgl5KNoulpSzgDaAAaftSFDsu5y23OF12m89USFZGNRhFNGejwcB8xjBoCWn8ND1MTy6XgJOcqvUVWnGq05SV6pL+iHMpNo8vZ+n2lBV6mROG6m8UyhaZWtc5jKwp6CIplVzCq/dfkhinwZAOJayY1sR9BIsgBFOVb4bSxphMsQD2tEwwD6VvWSchDKPtdI3vVDoF4A5Vv86EuxLOfJdPZmeKBd7qfhfrxODd9lLWBDFQAMyWeTF3ITQTGM1+0xQAnykASaC6VtN8ogFJqDDT0OwF7+8DJqnZfusbN2dFyvWNI4cOkRHqIRsdIEq6AbVUAMR9Ize0Sf6Ml6MD+Pb+JmMZoyp5wDNVSb7C7wPta4=</latexit> lD= DD· lp <latexit sha1_base64="+hWCyvoddyjixDhdrwMA9gMrRtU=">AAACGHicbVDLSsNAFJ34rPUVdelmsAiuaqKCboSCXbisYB/QhDCZTNqhk0mYmQgl5DPc+CtuXCjitjv/xkkbUVsPDJw5517uvcdPGJXKsj6NpeWV1bX1ykZ1c2t7Z9fc2+/IOBWYtHHMYtHzkSSMctJWVDHSSwRBkc9I1x/dFH73gQhJY36vxglxIzTgNKQYKS155qkTITX0w4x5zfz6+9P0smaeOziI1Y+fJXle9cyaVbemgIvELkkNlGh55sQJYpxGhCvMkJR920qUmyGhKGYkrzqpJAnCIzQgfU05ioh0s+lhOTzWSgDDWOjHFZyqvzsyFEk5jnxdWawp571C/M/rpyq8cjPKk1QRjmeDwpRBFcMiJRhQQbBiY00QFlTvCvEQCYSVzrIIwZ4/eZF0zur2ed26u6g1rDKOCjgER+AE2OASNMAtaIE2wOARPINX8GY8GS/Gu/ExK10yyp4D8AfG5AtrpKEx</latexit> TI <latexit sha1_base64="EwqXb61d9qJXvgceE4jHgyFQJyU=">AAAB9XicbVDLSgMxFL3js9ZX1aWbYBFclYwKuiy40V2FvqAdSybNtKGZzJBklDLMf7hxoYhb/8Wdf2OmnYW2HggczrmXe3L8WHBtMP52VlbX1jc2S1vl7Z3dvf3KwWFbR4mirEUjEamuTzQTXLKW4UawbqwYCX3BOv7kJvc7j0xpHsmmmcbMC8lI8oBTYqz00A+JGftB2hykd1k2qFRxDc+AlolbkCoUaAwqX/1hRJOQSUMF0brn4th4KVGGU8Gycj/RLCZ0QkasZ6kkIdNeOkudoVOrDFEQKfukQTP190ZKQq2noW8n85R60cvF/7xeYoJrL+UyTgyTdH4oSAQyEcorQEOuGDViagmhitusiI6JItTYosq2BHfxy8ukfV5zL2r4/rJax0UdJTiGEzgDF66gDrfQgBZQUPAMr/DmPDkvzrvzMR9dcYqdI/gD5/MH3saSsA==</latexit> TI <latexit sha1_base64="EwqXb61d9qJXvgceE4jHgyFQJyU=">AAAB9XicbVDLSgMxFL3js9ZX1aWbYBFclYwKuiy40V2FvqAdSybNtKGZzJBklDLMf7hxoYhb/8Wdf2OmnYW2HggczrmXe3L8WHBtMP52VlbX1jc2S1vl7Z3dvf3KwWFbR4mirEUjEamuTzQTXLKW4UawbqwYCX3BOv7kJvc7j0xpHsmmmcbMC8lI8oBTYqz00A+JGftB2hykd1k2qFRxDc+AlolbkCoUaAwqX/1hRJOQSUMF0brn4th4KVGGU8Gycj/RLCZ0QkasZ6kkIdNeOkudoVOrDFEQKfukQTP190ZKQq2noW8n85R60cvF/7xeYoJrL+UyTgyTdH4oSAQyEcorQEOuGDViagmhitusiI6JItTYosq2BHfxy8ukfV5zL2r4/rJax0UdJTiGEzgDF66gDrfQgBZQUPAMr/DmPDkvzrvzMR9dcYqdI/gD5/MH3saSsA==</latexit> lS= S· l <latexit sha1_base64="cQAJAtoYJ5Dbv2ZjBb9Y+yAmkFo=">AAACEnicbVDLSsNAFJ34rPUVdelmsAi6KYkKuhEKblxWah/QhDCZTNqhk5kwMxFKyDe48VfcuFDErSt3/o1JG0VbDwycOede7r3HjxlV2rI+jYXFpeWV1cpadX1jc2vb3NntKJFITNpYMCF7PlKEUU7ammpGerEkKPIZ6fqjq8Lv3hGpqOC3ehwTN0IDTkOKkc4lzzx2IqSHfpgyL21l2eX3t5U5OBD6x82qnlmz6tYEcJ7YJamBEk3P/HACgZOIcI0ZUqpvW7F2UyQ1xYxkVSdRJEZ4hAakn1OOIqLcdHJSBg9zJYChkPnjGk7U3x0pipQaR35eWayoZr1C/M/rJzq8cFPK40QTjqeDwoRBLWCRDwyoJFizcU4QljTfFeIhkgjrPMUiBHv25HnSOanbp3X75qzWsMo4KmAfHIAjYINz0ADXoAnaAIN78AiewYvxYDwZr8bbtHTBKHv2wB8Y71+vqJ6q</latexit> lP= P· lS <latexit sha1_base64="UgXNgpxHHfDlHNkRMkoPVbGWzYk=">AAACFnicbVDLSsNAFJ3UV62vqEs3g0VwY0lU0I1QcOMyon1AE8JkOmmHTjJhZiKUkK9w46+4caGIW3Hn3zhpI9TWAwNnzrmXe+8JEkalsqxvo7K0vLK6Vl2vbWxube+Yu3ttyVOBSQtzxkU3QJIwGpOWooqRbiIIigJGOsHouvA7D0RIyuN7NU6IF6FBTEOKkdKSb564EVLDIMyYnzl5fvX7dXIX97mace/yvOabdathTQAXiV2SOijh+OaX2+c4jUisMENS9mwrUV6GhKKYkbzmppIkCI/QgPQ0jVFEpJdNzsrhkVb6MORCv1jBiTrbkaFIynEU6MpiTTnvFeJ/Xi9V4aWX0ThJFYnxdFCYMqg4LDKCfSoIVmysCcKC6l0hHiKBsNJJFiHY8ycvkvZpwz5r2Lfn9aZVxlEFB+AQHAMbXIAmuAEOaAEMHsEzeAVvxpPxYrwbH9PSilH27IM/MD5/AAaYoHY=</latexit> dR= MR·y dG= MG·y dB= MB·y <latexit sha1_base64="cCjIntSEL9DH+slwG3occc0Q9Ms=">AAACWXichVFdS8MwFE2r01m/qnv0pTgEn0Y7BX0RxnyYL8IU9wFrGWmabmHpB0kqlNI/6YMg/hUfTLsKcxO8kHA451zuzYkbU8KFaX4o6tZ2bWe3vqftHxweHesnp0MeJQzhAYpoxMYu5JiSEA8EERSPY4Zh4FI8chf3hT56xYyTKHwRaYydAM5C4hMEhaSmemwHUMxdP/Om2XN+91jcNvIikf8IaW7b2oqrV7p6/7i6pau74dKmetNsmWUZm8CqQBNU1Z/qb7YXoSTAoUAUcj6xzFg4GWSCIIpzzU44jiFawBmeSBjCAHMnK5PJjQvJeIYfMXlCYZTsakcGA87TwJXOYkW+rhXkX9okEf6tk5EwTgQO0XKQn1BDREYRs+ERhpGgqQQQMSJ3NdAcMoiE/IwiBGv9yZtg2G5ZV63203WzY1Zx1MEZOAeXwAI3oAMeQB8MAALv4EupKTvKp6qodVVbWlWl6mmAX6U2vgHNF7c8</latexit>

Figure 3: Different steps involved in the development pipeline.

3.1

Photo-site selection

The selection is composed of k2neighboring blocks, each composed of 8 × 8 photo-sites with the surrounding border included since interpolated values inside a 8 × 8 block need photo-site values surrounding this block (see Figure 4). The result of flattening this matrix is the vector of the photo-sites samples denoted y∈N(8k+2)2.

(5)

Conference’17, July 2017, Washington, DC, USA Taburet, Bas, Fridrich & Sawaya

3.2

Demosaicking

The RAW cover image from the sensor first undergoes the demo-saicking process. To this end, we retrieve the interpolation kernels used by the linear demosaicking process to generate the full color image (i. e., a full set of color triples) from the CFA (see Figure 4).

It is possible to write the demosaicking operations as matrix multiplications. We define matrices MR, MG, MBof size (8k + 2)2× (8k + 2)2, such that the result of the matrix multiplication of the vector y with one of these matrices is the unfolded version of the corresponding color channel after demosaicking:

dR= MR· y, dG= MG· y, dB= MB· y. (6) For example, for the green channel, assuming that i ∈ N(8k+2)2 represents the photo-site index in y, there are two cases to compute the matrix MG.

(1) If index i corresponds to a Green photo-site on the Bayer CFA, this photo-site does not need color interpolation:

rowi(MG)= unfold © ­ ­ ­ ­ ­ ­ ­ « i . .. ... ... ... ... · · · 0 0 0 · · · · · · 0 1 0 · · · · · · 0 0 0 · · · ... ... ... ... ... ª ® ® ® ® ® ® ® ¬ i (7)

(2) If index i corresponds to a pixel which needs to be interpo-lated, then: rowi(MG)= unfold © ­ ­ ­ ­ ­ ­ ­ « i . .. ... ... ... ... · · · 0 0.25 0 · · · · · · 0.25 0 0.25 · · · · · · 0 0.25 0 · · · ... ... ... ... . .. ª ® ® ® ® ® ® ® ¬ i (8)

Note that for channels Red and Blue, we have to use four different convolution kernels to obtain the equations similar to equations (7) and (8): [1],       0.25 0 0.25 0 0 0 0.25 0 0.25       ,       0.5 0 0.5       and  0.5 0 0.5  .

3.3

Luminance averaging:

According to the standard ITU-R BT 601 for RGB color spaces, the relative luminance can be calculated as a linear combination of RGB components:

Y= 0.299 · dR+ 0.587 · dG+ 0.114 · dB. (9) In our case, this means that to perform color reconstruction and luminance averaging, we can define the demosaicking and averaging matrix DDas:

l= DD· y= (0.299 · MR+ 0.587 · MG+ 0.114 · MB) · y. (10) 1 + (3<latexit sha1_base64="NVweD1oBdjAW2756xuTk+walSy8=">AAAB9XicbVDLSgNBEOyNrxhfUY9eBoMQCYRdI5hjwIsHDxHMA5I1zE4myZDZBzO9SljyH148KOLVf/Hm3zhJ9qCJBQ1FVTfdXV4khUbb/rYya+sbm1vZ7dzO7t7+Qf7wqKnDWDHeYKEMVdujmksR8AYKlLwdKU59T/KWN76e+a1HrrQIg3ucRNz16TAQA8EoGunBKRUrXRQ+19XzktPLF+yyPQdZJU5KCpCi3st/dfshi30eIJNU645jR+gmVKFgkk9z3VjziLIxHfKOoQE1i9xkfvWUnBmlTwahMhUgmau/JxLqaz3xPdPpUxzpZW8m/ud1YhxU3UQEUYw8YItFg1gSDMksAtIXijOUE0MoU8LcStiIKsrQBJUzITjLL6+S5kXZqZTtu8tC7TaNIwsncApFcOAKanADdWgAAwXP8Apv1pP1Yr1bH4vWjJXOHMMfWJ8/KN+RBA==</latexit> ⇥ 8) + 1 1+ (3 ⇥ 8) + 1 <latexit sha1_base64="NVweD1oBdjAW2756xuTk+walSy8=">AAAB9XicbVDLSgNBEOyNrxhfUY9eBoMQCYRdI5hjwIsHDxHMA5I1zE4myZDZBzO9SljyH148KOLVf/Hm3zhJ9qCJBQ1FVTfdXV4khUbb/rYya+sbm1vZ7dzO7t7+Qf7wqKnDWDHeYKEMVdujmksR8AYKlLwdKU59T/KWN76e+a1HrrQIg3ucRNz16TAQA8EoGunBKRUrXRQ+19XzktPLF+yyPQdZJU5KCpCi3st/dfshi30eIJNU645jR+gmVKFgkk9z3VjziLIxHfKOoQE1i9xkfvWUnBmlTwahMhUgmau/JxLqaz3xPdPpUxzpZW8m/ud1YhxU3UQEUYw8YItFg1gSDMksAtIXijOUE0MoU8LcStiIKsrQBJUzITjLL6+S5kXZqZTtu8tC7TaNIwsncApFcOAKanADdWgAAwXP8Apv1pP1Yr1bH4vWjJXOHMMfWJ8/KN+RBA==</latexit> … … … (3 ⇥ 8 + 2) 2 <latexit sha1_base64="HGcCyfLhg1Y3w2GoKa+Nokr6z14=">AAAB+XicbVBNS8NAEJ3Ur1q/oh69LBahIpSkFeyx4MWDhwr2A9pYNttNu3SzCbubQgn9J148KOLVf+LNf+O2zUFbHww83pthZp4fc6a043xbuY3Nre2d/G5hb//g8Mg+PmmpKJGENknEI9nxsaKcCdrUTHPaiSXFoc9p2x/fzv32hErFIvGopzH1QjwULGAEayP1bbtU7WkWUlW7qlw+pZVZ3y46ZWcBtE7cjBQhQ6Nvf/UGEUlCKjThWKmu68TaS7HUjHA6K/QSRWNMxnhIu4YKbJZ56eLyGbowygAFkTQlNFqovydSHCo1DX3TGWI9UqveXPzP6yY6qHkpE3GiqSDLRUHCkY7QPAY0YJISzaeGYCKZuRWREZaYaBNWwYTgrr68TlqVslstOw/Xxfp9FkcezuAcSuDCDdThDhrQBAITeIZXeLNS68V6tz6WrTkrmzmFP7A+fwC/K5J2</latexit>

Figure 4: Unfolding of the photo-sites array fork = 3. Bayer patterns for the DCT blocks are surrounded by bold lines.

1 2 3

(a)

1 2 3

(b)

Figure 5: Fork = 3, scan order of the unfolded version of (a) lS, (b) lP

3.4

Selection of pixels of interest

As stated above, the surrounding edges of k × k blocks of samples have been included in order to take into account the convolution window during demosaicking. Once the demosaicking operations have been carried out, they can now be discarded. We consequently define a selection matrix S to perform this operation on the flattened version of the output l of the luminance averaging step such that : lS= S · l, (11) where S is composed of an arrangement of k · 64 × k · 64 identity matrices allowing us to select horizontal segments of 8k pixels. The resulting vector after the selection step is a (8k)2dimensional vector that we denote lS∈R(8k)2with S∈R(8k)2×(8k+2)2.

3.5

Permutation

As explained above, we inherently work with an arrangement of k × k blocks of photo-site values. However, the flattening operation induces by default a row scan and not a block-by-block scan. To this end, we define a permutation matrix P:

lP= P · lS, (12) where P is an arrangement of 8 × 8 photo-sites used to select hori-zontal segments of 8 pixels. This allows us to rearrange the vector lSinto lPas illustrated in Figure 5.

3.6

Vectorized DCT Transform

Given an 8 × 8 block in the spatial domain, Xs, its block 8 × 8 DCT transform denoted here as XDCTcan be written as the following matrix multiplication :

(6)

for Natural Steganography in JPEG Domain Conference’17, July 2017, Washington, DC, USA With : A=                 a a a a a a a a b d e д −д −e −d −b c f −f −c −c −f f c d −д −b −e e b д −d a −a −a a a −a −a a e −b д d −d −д b −e f −c c −f −f c −c f д −e d −b b −d e −д                 , and : [a,b, c,d, e, f ,д]t=

[cosπ4, cos16π, cosπ8, cos3π16, cos5π16, cos3π8 , cos7π16]t. It should be observed that the multiplication by A and At is due to the fact that the DCT transform is separable and processes the columns and rows independently.

In order to compute the covariance matrix of the spatial signal XS, we use vector notation by transforming the matrix XS∈R8×8 into a vector xS∈R64by concatenating the columns.

As a result, the 8×8 matrix A is transformed into a 64×64 matrix Avgiven by : Av=           A 0 . . . 0 0 A 0 ... .. . 0 . .. 0 0 · · · 0 A           . (13)

We also define a transpose operator as the matrix T ∈ R64×64 (see [4]) and the transpose operation Xt

Sis then equivalent to the multiplication T · xS, and the vector form of the DCT 8 × 8 block XDCTfinally becomes :

DCTv(xS)= Av· T · Av· T · xS= DT· xS, (14) where DT= Av· T · Av· T.

3.7

Covariance matrix computation

The matrix multiplication sequence can thus be written as tI= (DT· P · S · DD) · y. (15) Each matrix is thus a linear operator allowing us to calculate the vectorized DCT transform of k × k blocks from a sample of photo-sites. Therefore, only the y vector depends on the content of the RAW image. The matrix product DT·P·S·DDis content independent and can therefore be written as a single matrix.

The covariance matrix can now be calculated as follows: Σ = E[tI· ttI]= M · E[y · yt] · Mt, (16) with

M= DT· P · S · DD. (17)

4

RESULTS

In this section, we evaluate the empirical security of NS in the JPEG domain for images acquired with a color sensor. A total of 4800 512×512 images were obtain by partitionning into non-overlapping patches 100 RAW images acquired at ISO2using the Z-CAM-E1 camera [4] to build the covers subset. In a likewise way the pseudo-stego subset of 4800 512 × 512 images is built from 100 RAW images

acquired at ISO1and undergoes a switch [1] from ISO1= 100 to ISO2= 200. The parameters to perform the switch from ISO1to ISO2were (a,b) = (1.15, −1150). The empirical security is evalu-ated as the minimal total classification error probability under equal priors, PE= minPFA12(PFA+ PMD), with PF Aand PM Dstanding for the false-alarm and missed detection rates. The JPEG images are ste-ganalyzed with the DCTR feature set [10] and the low-complexity linear classifier [3].

The following embedding schemes are compared:

- Pseudo-embedding, simulated embedding is performed at the photo-site level with the stego signal generated using (2). These results can be considered as a baseline but do not correspond to any practical embedding scheme in the JPEG domain.

- 4 Lattices-scaling: The embedding pipeline uses the estimated covariance matrix to perform the sampling. The covariance matrix used is ˆΣ as explained in Section 2.2.

- 4 Lattices-analytic: The sampling mechanism is the same but we use the closed-form expression for the covariance matrix Σ, detailed in Section 3, to simulate the stego signal in the DCT domain.

- SI-UNIWARD: For comparison with the current state of the art, all images have also been embedded using SI-UNIWARD [11] with the embedding rate set to 1 bit per nzAC DCT coefficient which corresponds to the maximal payload of this embedding scheme.

The capacity of both 4-Lattice implementations ranges from approximately 1 bpnzAC at QF 75 to 2 bpnzAC at QF 100 (see Figure 11 of [16]). The results of these experiments are shown in Table 1. First, observe that there is no difference between generating the stego signal at the photo-sites or in the JPEG domain, which validates the fact that our statistical model in the JPEG domain is equivalent to the one at the photo-site level. Second, the closed-form of the covariance matrix provides a security gain w.r.t. the scheme proposed in [16], especially for high JPEG QFs. This is due to the fact that the covariance estimation proposed in [16] deals with blocks of constant photo-site values. This approximation is detrimental whenever ones wants to generate a high resolution stego signal. Note, however, that the computation of the covariance matrix is associated at a high computational burden since the Shur complement matrix needed to compute the conditional probability distribution has to be evaluated for each block and not only once as in [16]. The comparison with SI-UNIWARD shows that cost-based SI-embedding is more detectable than NS embedding.3

In Table 2 we evaluate the sensitivity of our methods w.r.t. other popular demosaicking schemes and compare the security of NS im-plemented with the closed-form of the covariance matrix and with scaled estimated covariance as proposed in [16]. Cover images are developed using a speficic demosaicking method and stego images are generated either assuming bilinear demosaicking (the value on the left) or with the scaled estimated covariance matrix [16] as also explained in Section 2.2. While the closed-form of the covariance matrix offers the best performance for bilinear demosaicking, it can-not be used to model other demosaicking schemes, in which case it is better to estimate the covariance matrix and scale it, especially for high QFs.

3Note that SI-UNIWARD only needs an uncompressed image while NS needs

(7)

Conference’17, July 2017, Washington, DC, USA Taburet, Bas, Fridrich & Sawaya JPEG Pseudo 4 Lattices 4 Lattices SI-Uniward

QF embedding scaling [16] analytical 1 bpnzac 100 40.2 13.9 42.9 0.0

95 40.9 30.3 41.2 0.4 85 41.9 39.8 41.2 12.3 75 41.3 40.4 41.6 24.8 Table 1: Empirical security (PEin %) for different quality fac-tors and embedding strategies on E1Base with bilinear de-mosaicking. DCTR features combined with regularized lin-ear classifier are used for steganalysis.

JPEG QF Bilinear VNG DCB AAHD 100 42.9 / 13.9 0.0 / 0.0 0.0 / 0.1 0.0 / 0.0 95 41.2 / 30.3 0.1 / 22.7 0.1 / 4.5 0.1/ 3.4 85 41.2 / 39.8 0.4 / 36.9 0.6 / 32.6 1.1/ 25.4 75 41.6 / 40.4 1.5 / 40.9 3.9 / 39.8 1.1/ 35.7 Table 2: Empirical security (PE (%) and sensitivity w.r.t the demosaicking algorithm used to develop cover images us-ing analytic covariance matrix computed for bilinear demo-saicking (left), or using scaled estimated covariance matrix (right) [16].

5

CONCLUSIONS AND FURTHER WORKS

This paper is an extension of the former works done on Natural Steganography in the JPEG domain [4, 16] and on color sensors. We have shown that for the bilinear development pipeline, it is possible to derive an explicit form of the covariance matrix of the stego signal in the DCT domain. Similar derivations should be possible for other linear demosaicking algorithms such as VNG. When used for embedding, the closed-form of the covariance matrix enables us to simulate embedding in the JPEG domain with the same empirical security as in the photo-site domain. However, when the demosaicking process is not linear, it is then better to estimate the covariance matrix using Monte Carlo sampling as in [16] and scale it according to (3) than to use the closed-form derived here.

Future work will focus first on finding non-Gaussian models that match the distributions of non-linear demosaicking algorithms, such as DCB or AAHD. This can be done for example by contin-uing to use a covariance matrix to model dependencies but also by modifying marginals to fit arbitrary distributions. A second line of research concerns applications of the derived closed-form covariance matrix Σ to steganography. It may allow considering proper synchronization of embedding changes in classical, cost-based steganography similar to [12].

ACKNOWLEDGMENTS

This work has been funded in part by the French National Re-search Agency (ANR-18-ASTR-0009), ALASKA project: https:// alaska.utt.fr, and by the French ANR DEFALS program (ANR-16-DEFA-0003). The work on this paper was also supported by NSF grant No. 1561446.

REFERENCES

[1] Patrick Bas. 2016. Steganography via Cover-Source Switching. IEEE Workshop on Information Forensics and Security (WIFS).

[2] Patrick Bas. 2017. An embedding mechanism for Natural Steganography after down-sampling. IEEE ICASSP.

[3] Rémi Cogranne, Vahid Sedighi, Jessica Fridrich, and Tomáš Pevn`y. 2015. Is ensemble classifier needed for steganalysis in high-dimensional feature spaces?. In Information Forensics and Security (WIFS), 2015 IEEE International Workshop on. IEEE, 1–6.

[4] Tomáš Denemark, Patrick Bas, and Jessica Fridrich. 2018. Natural Steganography in JPEG Compressed Images. In Electronic Imaging. San Francisco, United States. [5] Tomas Denemark and Jessica Fridrich. 2015. Side-informed steganography with

additive distortion. In Information Forensics and Security (WIFS), 2015 IEEE Inter-national Workshop on. IEEE, 1–6.

[6] T. Denemark and J. Fridrich. 2017. Steganography with Multiple JPEG Images of the Same Scene. 12, 10 (October 2017), 2308–2319.

[7] Tomas Filler, Jan Judas, and Jessica Fridrich. 2011. Minimizing additive distortion in steganography using syndrome-trellis codes. Information Forensics and Security, IEEE Transactions on6, 3 (2011), 920–935.

[8] Jessica Fridrich. 2009. Steganography in digital media: principles, algorithms, and applications. Cambridge University Press.

[9] Jessica Fridrich, Miroslav Goljan, and David Soukal. 2004. Perturbed quantization steganography with wet paper codes. In Proceedings of the 2004 workshop on Multimedia and security. ACM, 4–15.

[10] Vojtěch Holub and Jessica Fridrich. 2015. Low-complexity features for JPEG steganalysis using undecimated DCT. IEEE Transactions on Information Forensics and Security10, 2 (2015), 219–228.

[11] Vojtěch Holub, Jessica Fridrich, and Tomáš Denemark. 2014. Universal distor-tion funcdistor-tion for steganography in an arbitrary domain. EURASIP Journal on Information Security2014, 1 (2014), 1–13.

[12] Weixiang Li, Weiming Zhang, Kejiang Chen, Wenbo Zhou, and Nenghai Yu. 2018. Defining Joint Distortion for JPEG Steganography. In Proceedings of the 6th ACM Workshop on Information Hiding and Multimedia Security. ACM, 5–16. [13] A. Papoulis and U. Pillai. 2002. Probability, Random Variables and Stochastic

Processes. Mac Graw Hill.

[14] P. Sallee. 2003. Model-Based Steganography. In International Workshop on Digital Watermarking (IWDW), LNCS, Vol. 2.

[15] Vahid Sedighi, Rémi Cogranne, and Jessica Fridrich. 2016. Content-Adaptive Steganography by Minimizing Statistical Detectability. Information Forensics and Security, IEEE Transactions on11, 2 (2016), 221–234.

[16] Théo Taburet, Patrick Bas, Wadih Sawaya, and Jessica Fridrich. 2019. A Natural Steganography Embedding Scheme Dedicated to Color Sensors in the JPEG Domain. In Electronic Imaging. Burlingame, United States.

[17] Z. Wang, Z. Qian, X. Zhang, M. Yang, and D. Ye. 2018. On Improving Distortion Functions for JPEG Steganography. IEEE Access 6 (2018), 74917–74930. https: //doi.org/10.1109/ACCESS.2018.2884198

[18] Kuan Yang, Kejiang Chen, Weiming Zhang, and Nenghai Yu. 2018. Provably Se-cure Generative Steganography Based on Autoregressive Model. In International Workshop on Digital Watermarking. Springer, 55–68.

[19] Zhongliang Yang, Xiaoqing Guo, Ziming Chen, Yongfeng Huang, and Yu-Jin Zhang. 2018. RNN-Stega: Linguistic Steganography Based on Recurrent Neural Networks. IEEE Transactions on Information Forensics and Security (2018).

Figure

Figure 1: Overview of the embedding scheme presented in [16].
Figure 2: The four lattices Λ 1 , Λ 2 , Λ 3 and Λ 4 , used for embed- embed-ding.
Figure 4: Unfolding of the photo-sites array for k = 3. Bayer patterns for the DCT blocks are surrounded by bold lines.
Table 1: Empirical security (P E in %) for different quality fac- fac-tors and embedding strategies on E1Base with bilinear  de-mosaicking

Références

Documents relatifs

HybriLIT includes GOVORUN supercomputer and education and testing polygon; its platform is based on the latest computation architectures (processors; co-processors;

Thus each mathematical symbol (G, L,. .) has a single definite meaning inside each exercise, but might have different meanings from an exercise to another.. It is not required to

In this paper, we propose an alternative to reparameteriza- tion based on a convex optimization formulation. Given any parameterized curve, our algorithm returns the closest curve

Abstract—This paper proposes a new steganographic scheme relying on the principle of “cover-source switching”, the key idea being that the embedding should switch from one cover-

(Et, au demeurant, « pourquoi ? ») À titre d’artiste et de sociologue amatrice autoproclamée, Ann Hirsch, alias Caroline (Scandalishious.com), le demande ouvertement : «

To bring this idea closer to a practical em- bedding method, in this paper we embed the message in quantized DCT coefficients of a JPEG file by adding independent realiza- tions of

It is found that chipless RFID is capable of discriminating practically two different tags affected by the similar geometrical variations that are inherent to the fabrication

Among various types of optical aberrations from which a mosaic image may suffer, chromatic aberration (CA) influences the spatial and spectral correlation through the artefacts such