• Aucun résultat trouvé

Additive and multiplicative notions of leakage, and their capacities

N/A
N/A
Protected

Academic year: 2021

Partager "Additive and multiplicative notions of leakage, and their capacities"

Copied!
16
0
0

Texte intégral

(1)

HAL Id: hal-00989462

https://hal.inria.fr/hal-00989462

Submitted on 11 May 2014

HAL

is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire

HAL, est

destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Additive and multiplicative notions of leakage, and their capacities

Mário S. Alvim, Konstantinos Chatzikokolakis, Annabelle Mciver, Carroll Morgan, Catuscia Palamidessi, Geoffrey Smith

To cite this version:

Mário S. Alvim, Konstantinos Chatzikokolakis, Annabelle Mciver, Carroll Morgan, Catuscia Palamidessi, et al.. Additive and multiplicative notions of leakage, and their capacities. 27th Computer Security Foundations Symposium (CSF 2014), Jul 2014, Vienna, Austria. pp.308–322,

�10.1109/CSF.2014.29�. �hal-00989462�

(2)

Additive and multiplicative notions of leakage, and their capacities

M´ario S. Alvim Konstantinos Chatzikokolakis Annabelle McIver Carroll Morgan§ Catuscia Palamidessi Geoffrey Smith

Department of Computer Science Universidade Federal de Minas Gerais

Belo Horizonte

INRIA, CNRS and LIX Ecole Polytechnique´

Paris

Department of Computing Macquarie University

Sydney

§ School of Computer Science & Engineering University of New South Wales and NICTA

Sydney

School of Computing & Information Sciences Florida International University

Miami Abstract—Protecting sensitive information from improper

disclosure is a fundamental security goal. It is complicated, and difficult to achieve, often because of unavoidable or even unpredictable operating conditions that can lead to breaches in planned security defences. An attractive approach is to frame the goal as a quantitative problem, and then to design methods that measure system vulnerabilities in terms of the amount of information they leak. A consequence is that the precise operating conditions, and assumptions about prior knowledge, can play a crucial role in assessing the severity of any measured vunerability.

We develop this theme by concentrating on vulnerability measures that arerobustin the sense of allowing general leakage bounds to be placed on a program, bounds that apply whatever its operating conditions and whatever the prior knowledge might be. In particular we propose a theory of channel capacity, generalising the Shannon capacity of information theory, that can apply both to additive- and to multiplicative forms of a recently- proposed measure known as g-leakage. Further, we explore the computational aspects of calculating these (new) capacities: one of these scenarios can be solved efficiently by expressing it as a Kantorovich distance, but another turns out to be NP-complete.

We also find capacity bounds for arbitrary correlations with data not directly accessed by the channel, as in the scenario of Dalenius’s Desideratum.

Keywords—Quantitative information flow, channel capacity, confidentiality, Dalenius Desideratum.

I. Introduction

Protecting sensitive information from improper disclosure is a fundamental security goal, but one that is clearly not being achieved well in today’s cyber infrastructure. The issue is complicated by the fact that some leakage of sensitive information is often unavoidable in practice, due either to system functionality (e.g. a statistical database must by design reveal information derived from the database entries, even if those entries themselves should be confidential) or due to side-channels (e.g. it is difficult to prevent running time or power consumption from depending on secrets). For this reason, it is attractive to approach the problem quantitatively, and indeed there has been good progress in recent years in the area of quantitative information flow, including works with a foundational focus [1]–[11], works aimed at the development

of verification techniques [12]–[19], and works analyzing the leakage in real-system vulnerabilities [20]–[22].

One important quantitative theme has been the development of leakage measures with strong operational significance for confidentiality, so that the amount of information leaked is associated with the constraints and rewards of real attacks.

In this respect, notable measures include min-entropy leakage [6] and g-leakage [11]. Assuming that an adversary’s prior knowledge about the secret is represented as a probability distribution π, min-entropy leakage is based on the ratio of the secret’s posterior- and priorvulnerabilityto being guessed correctly in one try by the adversary. One-try vulnerability is clearly a basic and important measure of confidentiality, but it should be recognized that it is not appropriate in all situations.

For this reason,g-leakage generalizes vulnerability with again function g that models the operational scenario, allowing us to consider situations in which the adversary gains e.g. by guessing the secret partially, approximately, or within ktries.

It can also model scenarios where there is apenaltyfor making incorrect guesses.

A second important theme concernsrobustness. The leak- age caused by a channelC depends on both the prior distribu- tion πand the particular leakage measure used (e.g. Shannon entropy, guessing entropy, min-entropy,g-leakage). But what if we do not knowπ, or don’t know the leakage measure, or don’t know either? Then we might worry that our conclusions about information leakage lack robustness. Precisely for this reason, there has been considerable interest in achieving results that are less sensitive to the particular assumptions made of prior or entropy-measure.Capacityis the maximum leakage over all prior distributions—and its utility is that when a channel has small capacity, its leakage is small, no matter what the prior may be. It is also interesting to establish capacity relationships between different leakage measures. For instance, theMiracle Theoremof [11] shows thatmin-capacity(the maximum min- entropy leakage over all priors) is an upper bound ong-leakage foreverygain functiong—this means that a channel with small min-capacity has smallg-leakage, no matter what prior or gain function is used. A second approach to robustness concerns comparisons between channels. It has recently been shown in the Coriaceous Theoremof [23] that a channelA never leaks

(3)

more than a channelB, on any priorπand gain functiong, if and only if A secures Bin the sense that that A is equivalent to the combination of Bwith some “post-processing”R.

This paper builds on these successes of the g-leakage model, achieving advances in both operational significance and robustness.

With respect to operational significance, recall that (as mentioned above) standard g-leakage is based on the ratio of the posterior and prior g-vulnerabilities. But there are situations where it seems more significant to consider instead the difference of the vulnerabilities. For instance, a system whose prior- and posterior vulnerabilities are 2−1000and 2−700 respectively has a huge multiplicative increase of 2300, but its additive increase is less than 2−700, which might well be considered negligible. Moreover,g-vulnerabilities can be used to model the economic value to the adversary of the prior- and posterior situations; here too it seems more significant to consider their difference rather than their ratio. Finally, we discover that both guessing-entropy and Shannon-entropy leakage can be expressed as additive g-leakages. For these reasons, in this paper we develop the theory of additive g- leakage, complementing the existing multiplicative theory.

With respect to robustness, this paper presents a systematic study of g-capacity. Since the g-leakage of a channel C de- pends on both the priorπand the gain functiong, we consider several versions of capacity: we can fix g and universally quantify π, or we can fix π and universally quantify g, or we can universally quantify both π and g. By considering both multiplicative- and additive leakage, we get a total of six capacity scenarios. We also consider a seventh capacity scenario related to the famous ad-omnia privacy desideratum proposed by Dalenius [24] for statistical databases, which states that nothing about an individual should be learnable with the database that could not equally well have been learned without it. As shown by Dwork [25], this desideratum is too strong to be achievable: for any useful database there is always some prior knowledge the adversary can use to induce a leakage of information. Analogously, we show here that g- leakage can quantify the leakage about a secretXcaused by an apparently unrelated channelCfromYtoZwhen the adversary knows some correlationsbetweenX andY.

Beyond our systematic study, we make novel and signifi- cant contributions in four of those seven capacity scenarios:

• Given a priorπ, we show that a gain functiongπ max- imizing the multiplicative leakage can be computed efficiently. (§VI-B)

• For the specific gain function gid, we show that the maximum additive leakage over all priors is NP- complete. (§VII-A)

• Given a priorπ, we show that the maximum additive leakage over all gain functions can be computed in lineartime, because this leakage can be expressed as aKantorovich distance. (§VII-B)

• We show that the “Dalenius” leakage of X caused by a channelC fromY toZ isbounded by the capacity of C, regardless of the correlations that may exist betweenX andY. (§IX)

The rest of this paper is organized as follows. Section II gives preliminaries, Section III motivates additive g-leakage, and Section IV presents useful algebraic properties for gain functions. Section V then defines the seven capacity scenarios that we consider. Sections VI and VII present our results on the multiplicative- and additive scenarios respectively, with Section VIII giving some examples. Section IX presents our

“Dalenius” bounds, proven using new results about the additive capacity of channel cascades. Finally, Sections X and XI discuss related work and conclude.

II. Preliminaries A. Basic notations

We begin by recalling the basic definitions of information- theoretic channels [26]. Achannelis a triple (X,Y,C) where X andY are finite sets (typically of secret input values and observable output values resp.) and C is an |X|×|Y| channel matrixwhose entries are between 0 and 1 and whose rows each sum to 1. Typically we use upper-case Roman letters (likeC) for channels, and calligraphic letters (like X,Y) for the sets over which they operate. We write Cx,y for the element ofC at row xinX and column yinY. For the (entire) row xwe writeCx,−, and for columnywe writeC−,y.

The valueCx,y is the conditional probability of output y’s being produced byC from inputx. A channel is deterministic just when all its elements are either 0 or 1, implying that each input row contains a single 1 which identifies that input’s unique corresponding output.

Typically we assume knowledge of a prior distribution π on X, using lower-case Greek letters for simple distributions like that; by analogy with Cx,y indexing a matrix, as above, we write πx for π at x, thus the probability that π gives to x. For more complex distributions we use upper-case Greek letters;1for example thejoint distributiononX×Ydetermined by input distribution π and channel C is typically written Πx,yxCx,y; when Π is understood, we can use a Π-implicit convention to write that as just p(x,y). The jointly distributed random variables X,Y havemarginal probabilities that (again Π-implicitly) we write p(x)=P

yp(x,y) (which of course is justπxagain) and p(y)=P

xp(x,y); occasionally we also write X,Y informally for “the (unnamed) distribution onX,Y” resp.

Theconditional probabilitiesare then given byp(y|x)=p(x,y)/p(x)

(if p(x) is non-zero) andp(x|y)=p(x,y)/p(y) (if p(y) is non-zero).

Note that Π is the unique joint distribution that recovers π andC, in that p(x)=πx and p(y|x)=Cx,y (if p(x) is non-zero).

When necessary to avoid ambiguity, we write theseΠ-implicit distributions with subscripts, e.g. pX or pXY or pY.

For example given |X|=3 and|Y|=4, we could consider a channelC (at left) which, when applied to (the uniform) prior π=(1/3,1/3,1/3) produces the joint matrixΠ(at right):

C y1 y2 y3 y4

x1 1 0 0 0

x2 0 1/2 1/4 1/4

x3 1/2 1/3 1/6 0

−→π

Π y1 y2 y3 y4

x1 1/3 0 0 0 x2 0 1/6 1/12 1/12

x3 1/6 1/9 1/18 0 (1) By summing Π’s columns we get theY-marginal distribution pY = (1/2,5/18,5/36,1/12), and by normalizing those columns

1Where the upper-case Greek- and Roman letters are the same, we will write the Greek in sans-serif: thus upper-caseτwould beT.

(4)

we obtain the four posterior distributions pX|y1 = (2/3,0,1/3), pX|y2=(0,3/5,2/5), pX|y3 =(0,3/5,2/5) and pX|y4=(0,1,0).

We use D() to construct the set of distributions on a set, so that DX is typically the set of distributions on the input andDY is the set of distributions on the output. For a point- or equivalently Dirac distribution assigning probability 1 to some xinX(and thus 0 to all other elements ofX) we write [x]. We write⌈π⌉for the support of distributionπ, that is the (sub-)set of elements where it is not zero.2 For the sum of a (sub-)distributionπwe write justΣπ; using that we could write more succinctly p(x)=PΠx,− and p(y)=PΠ−,y.

For theexpected value over a distribution π:DX of some random variable F:X→R, whereR is usually the realsR but more generally can be a vector space, we writeEπF; thus in the discrete case we haveEπF=P

xπxF(x). IfXitself is a vector space, then we abbreviate Eπ(id) by just Eπ, the “average” of the distribution πonX.

Leakagemeasures can be defined based on various entropy- like functions of the prior distribution π and the posterior distributions pX|y, together with their probabilities p(y). Here are some examples.

Shannon leakageis based on the Shannon entropy H[π]=

−P

xπxlgπx [27] of the prior distribution3 and the expected Shannon entropy of the posterior distributions H[π,C] = P

yp(y)H[pX|y]. It is their differenceH[π]−H[π,C] that equals the mutual information I(π,C).4

Guessing-entropy leakageis based on the guessing entropy [28] of the prior distribution, that isG[π]=P

iiπxi where the i-indexing ofX is in non-increasing probability order, and on the expected guessing entropy of the posterior distributions G[π,C]=P

yp(y)G[pX|y]. The guessing-entropy leakage then is the differenceG[π]G[π,C].

The operational significance of both Shannon- and guessing entropy can be stated in terms of the expected number of brute- force guesses an adversary would need to find the secret.5

Min-entropy leakage[6] is based on the priorvulnerability V[π]=maxxπx, the probability of the secret’s being guessed in one try, and the expected vulnerability of the posterior distributions V[π,C]=P

yp(y)V[pX|y].

The min-entropy leakage L(π,C) is the logarithm of the ratio of the posterior- and prior vulnerabilities:

L(π,C) = lg(V[π,C]/V[π]). (2) Note that vulnerability implicitly assumes an operational sce- nario in which the adversary gains only by guessing the secret exactly, and in one try.

For this reason,g-leakage[11] generalizes vulnerability to incorporate a gain function g, the choice of which allows the modelling of differing operational scenarios. In each scenario, there will be some set Wofguessesthat the adversary could

2Note that the characteristic function of the support set is indeed the ceiling of the distribution as a function.

3We write lg for log2 (Knuth).

4The more usual notation for these quantities isH(X),H(X|Y), andI(X;Y).

We explain at the end of§II why we use brackets [·].

5For Shannon entropy, this follows from a result by Massey [28].

make about the secret, and for any guesswand secret x, there will be some gain g(w,x) that the adversary gets by having chosen w when the secret’s actual value was x; gains range from 0, when choosingwhas no value at all, to 1, whenwis ideal. (In§III-A, and Def. 3 in§V this is generalised, however.) Formallyg:W×X →[0,1], whereWis a finite set. Given a gain function g, the prior g-vulnerability is defined as the maximum expected gain over all possible guesses:

Vg[π] = max

w

X

x

πxg(w,x). (3) The posterior g-vulnerability and the g-leakage are then de- fined as for min-entropy leakage above, so that we have

Vg[π,C] = X

y

p(y)Vg[pX|y], (4) and thenLg(π,C)=lg(Vg[π,C]/Vg[π]).

Note that the identity gain function gid, which gives gain 1 for guessing the secret correctly and 0 otherwise, makes g-leakage coincide with min-entropy leakage, which latter is thus a special case. But gain functions can do much more. As explained in [11], they can model a wide variety of practical operational scenarios, including those where the adversary benefits from guessing a value closeto the secret, guessing a partof the secret, guessing apropertyof the secret or guessing the secret within some bounded number of tries. They can also model scenarios where there is apenaltyfor incorrect guesses.

We now recall some previous results aboutcapacity.Shan- non capacityandmin-capacityare, respectively, the maximum Shannon leakage and maximum min-entropy leakage, over all priors. The min-capacity of C, denoted ML(C), is always realized on the uniform prior and it is equal to the logarithm of the sum of the column maximums of C [20], [29]. Min- capacity is an upper bound on Shannon capacity [11], and the two coincide on deterministic channels [30]. Finally, the Miracle Theorem [11] shows that min-capacity is an upper bound ong-leakage, foreveryprior πand gain functiong.

B. Hyper-distributions and their notations

A more abstract view of the action of a channel on a prior (e.g. more abstract than producing a joint distribution) is that it produces a distribution of posteriors. Following [7], [23], we call thosehyper-distributions, for brevity writing just “hyper”.

A hyper on the input space X is of type D2X. Consider a prior π on X and a channel C:X→Y which, as usual, determines a joint distribution Π:D(X×Y). As above, with Π understood we have a distribution pY on the outputs and, for each y, a corresponding posterior distribution pX|y on the inputs. If instead of considering pY to be a distribution on Y we consider it to be a distribution on the corresponding posteriors (i.e. on the normalised columns of Π themselves, rather than on theirY-labels), then we have a hyper which we write [π,C]. For the example at (1) above, that hyper would assign probabilities (1/2,15/36,1/12) to the posteriors (2/3,0,1/3), (0,3/5,2/5), and (0,1,0), respectively.6

6There might be fewer posteriors in the support of hyper [π,C] than there are columns in the joint distributionΠfrom which it derives, because if several columns ofΠnormalise to the same posterior then the hyper will automatically coalesce them [23]: columnsy2andy3were coalesced in this case.

(5)

Because hyper [π,C] is of type D2X, that is DV where V=DXis a vector space, we can just average that hyper itself, in effect “collapsing” it; and that gives the original prior again.

That is we have the convenient identityE[π,C]=π.

A special case of a hyper is thepointhyper on a particular (simple) distribution, say π: we write that as [π], as stated above for point distributions generally and of which this is just a special case. In this way we can regardg-vulnerabilityVgas acting on hypers in all cases. In particular Vg[π] viewed this way, i.e. as a functionVg of typeD2X→Rapplied to a (point) hyper, remains equal to maxwP

xπxg(w,x) as we defined it above.

III. AdditiveLeakage

Additive leakage is thedifferencebetween the posterior and prior vulnerabilities, and we write it

L+g(π,C) = Vg[π,C]Vg[π] .

Note that the additive version of min-entropy leakage (i.e. with respect to gid) was considered previously in [29] and [15].

(Standard g-leakage is defined multiplicatively, based on the ratio of the posterior and priorg-vulnerabilities.7)

We observe that calculating leakage additively rather than multiplicatively will not have any impact on chan- nels’ leakage ordering, since the additive leakage or- dering L+g(π,A)≤ L+g(π,B) and its multiplicative version Lg(π,A)≤ Lg(π,B) are both equivalent toVg[π,A]Vg[π,B]

(provided that Vg[π]>0 in the multiplicative case). But we believe that additive g-leakage is interesting and important in several respects, explained below.

A. Operational significance

First, there are a number of situations where additive leakage may have greater operational significance than does multiplicative leakage. To illustrate, consider a secret array X of uniformly distributed 10-bit passwords for 1,000 users. The following probabilistic channel (it is (Ex1) from [11]) reveals somerandomly-chosen user’s password:

u← {? 0..999}; Y =(u,X[u]);

When analyzed using min-entropy leakage, it leaks 10 bits, since it increases the vulnerability from 2−10000 to 2−9990, a factor of 210. But this analysis is inadequate if we are concerned about the loss ofanyuser’s password, and not just about the loss of theentiredatabase. Hence [11] defines a gain functiong where

W = {(u,x)|0≤u≤999 and 0≤x≤1023} and

g((u,x),X) =

( 1, if X[u]=x 0, otherwise.

Note that this g gives the adversary a gain of 1 for correctly guessing the password ofanyuser. When analyzed usingg, the channel now increases the g-vulnerability from 2−10 to 1. But

7Theg-leakage is thelogarithmof this ratio, which just changes the scale.

this is again a factor of 210, which means that the g-leakage is 10 bits, exactly the same as was the min-entropy leakage!

The point is that if we consider only the ratios of the vulnerabilities, an increase from 2−10000 to 2−9990 is exactly equivalent to an increase from 2−10 to 1. But, additively, the former increase is less than 2−9990, a completely negligible quantity, while the latter is about 0.999, a huge difference.8

Another important motivation for considering additive leak- age is the possibility of viewing vulnerabilities in economic terms. That is, we could define a gain function g so that the prior and posterior vulnerabilities would model the economic value to the adversary of its prior- and posterior knowledge about the secret. In this case, the additive leakage would represent the expected monetary gain from carrying out the attack represented by the channelC. Since making the attack might involve costs to the adversary, costs that he can predict, the additive leakage would then be crucial in his decision of whether the attack makes sense economically.

A technical issue to be considered concerns the range of gain functions. In [11], [23], gain functions are required to return values in the range [0,1]. Note that in the multiplicative context, negative vulnerabilities would cause mathematical problems, and merely allowing a larger range [0,a] would make no difference. In the additive context, however, it would make good sense to allow a wider range of gain values—

large values could represent large economic gains, and negative values could represent losses(as in the gain functiongtiger of [11]).9 We will say more about this issue in Section V below.

B. Expressiveness

Additive leakage is also interesting because some important existing leakages are special cases of additiveg-leakages.

First, we find that we can express guessing entropy as a negative g-vulnerability. If we let W be the set of all permutations f of X, we can then define the gain function

g(f,x) = −i

whereiis the uniqueindexofxwithin permutation f, assumed to range from 1 to n. (Note that this requires extending the allowed range of gain functions, sinceghas range in [−n,−1].) With this definition, we find thatVg[π]=−G[π] andVg[π,C]=

−G[π,C], which means that the additive g-leakage is exactly the guessing-entropy leakage:

L+g(π,C) = Vg[π,C]−Vg[π] = G[π]−G[π,C] . We might of course wonder whether guessing-entropy leakage could also be expressed as a multiplicative g-leakage, but we suspect that in general this is not possible.

Finally, we can even express Shannon leakage (mutual information) as an additive g-leakage, although this does require further extension.

8On the other hand, increasing vulnerability from 0.000001 to 0.300001 is far more significant than increasing it from 0.4 to 0.7, even though both give additive leakage of 0.3. Perhaps we could say, roughly speaking, that leakage is ‘significant’ only ifboththe multiplicative- and additive leakage are ‘large’.

9Theorem 4.1 of [11] says that we always haveVg(π,C)Vg(π). It remains true even if gain values are allowed to be negative, which means that additive leakage will always be non-negative.

(6)

To do this, we let W be the (uncountably infinite) set of all probability distributionsω onX, and we defineg by

g(ω,x) = lgωx .

Sinceωxis in [0,1], this gain function has range [−∞,0]. The g-vulnerability associated with this gain function is

Vg[π] = sup

ω

X

x

πxlgωx .

It turns out that Vg is realized when ω=π, thanks to Gibbs’

inequality [31], which implies that Vg[π] = −H[π]. This implies that the additive g-leakage for this g is exactly the Shannon leakage.

Using continuity arguments, we could restrict to acount- ablyinfinite Wand a gain function with range (−∞,0]. Still, W needs to be infinite andg unbounded in order to capture Shannon leakage. Although such gain functions are interesting from the point of view of expressiveness, in this paper we restrict to finite sets of guessesW.

IV. Gain-function algebra

Using gain functions to analyse leakage properties of chan- nels can be simplified by establishing some general algebraic properties. Fix a channelC, priorπ and gain function g. We define the following specialg-functions, fork:R:

g×k(w,x) = g(w,x)×k g+k(w,x) = g(w,x)+k

g+k@x(w,x) = g(w,x)+(kif x=xelse 0). (5) (Notice that in this section we are allowing completely unre- stricted gain functionsW×X →R, for some finite W.)

These new gain functions satisfy the following properties, which we prove in App. A:

x-Shift Vg+k@x[π,C] = Vg[π,C]+x Global Shift Vg+k[π,C] = Vg[π,C]+k Scale Vg×k[π,C] = k×Vg[π,C],

for k≥0

(6)

In consequence, we have the following invariance results for leakages, simple corollaries of the above equalities:

Lemma 1: Scaling invariance For fixed gain function g and channel C, multiplicative leakage is invariant under positive Scale.

Lemma 2: Shifting invariance For fixed gain functiong and channel C, additive leakage is invariant under Shift (of either kind).

V. Channel capacity

Shannon’s original definition of channel capacity uses mutual information as the measure of information flow and maximises over all priors. The classic significance of the value obtained is that it provides a tight bound on the amount of information that can be transmitted reliably over the channel.

In the security context, we look instead for bounds that tell us something about the “level of secrecy” of information, formulated for us in terms of relative (multiplicative) or comparative (additive) increases in g-vulnerability.

A useful notion of g-capacity would provide a robust measure of leakage that allowed meaningful predictions about how an attacker could use the channel for his gain. For robustness, we can consider universally quantifying over the prior π, over the gain function g, or over both. In quantifying over the priorπwe are acknowledging that, in many situations, the prior is unknown and the assumption that it is uniform is not reasonable. (For example, large-scale studies have shown that password selection is not at all uniform [32].) In quan- tifying over the gain function g we are acknowledging that we might not know the value to the adversary of different sorts of partial information about the secret, neither now nor even in the future. (In medical research, for instance, new correlations between diet and susceptibility to disease are constantly discovered: although knowing someone’s favorite dish might be relatively harmless today, it might not be so in the future.)

Since we can universally quantify over the prior π, over the gain function g, or over both, and since we are interested in both multiplicative- and additive g-leakage, we arrive at a total of six scenarios, which we write as follows:

Lg(∀,C),L(π,C),L(∀,C),L+g(∀,C),L+(π,C),L+(∀,C).

Each of them is of independent interest, and we define them precisely below.

For those scenarios that quantify over the gain functiong, a first issue to be resolved is the set of gain functions that we allow. In previous work [11], [23], a gain function is required to have type W × X →[0,1] for some finite set W. But, as motivated in Section III, with additive leakage we may wish to allow gain functions with a range different from [0,1]. Can we find a suitable set of gain functions to use in both multiplicative and additive g-capacity?

Note first that negative gain values are mathematically problematic for multiplicative leakage. But they are inessential for additive leakage, since negative gain values can be elim- inated by shifting (g is bounded sinceW is assumed finite), with no effect on additive leakage (Lemma 2). We therefore restrict our attention to non-negative gain functions.

Next consider an unbounded range [0,∞). Note that al- though any individual g is bounded (since W is finite), still an unbounded range allows to choose arbitrarily large gain functions when quantifying over them. For the multiplicative case this makes no difference, since anyg(since it’s bounded) can be scaled down to [0,1] without affecting the leakage (Lemma 2). On the other hand, scaling does affect the additive leakage, and the unbounded range of gcan cause the additive capacities L+(π,C), L+(∀,C) to be ill defined. Instead of bounding the range of g, we adopt the more permissive and mathematically attractive approach (for reasons explained in

§VII) to restrict to 1-spanning gain functions, whose gain values, for a given guess w, span a range of size at most 1:

Definition 3: 1-spanning A gain function g is 1- spanning if|g(w,x)g(w,x)| ≤1 for all x,x:X,w:W. We writeG1Xto denote the set of non-negative, 1-spanning gain functions onX. When we formally define the differentg- capacities below, this is the set of gain functions that we allow.

Note that bounded range [0,k] gain functions are a special case: they are scalar multiples of 1-spanning gain functions.

(7)

A. Multiplicative capacities defined

Recall that DX is the set of distributions over a finite (and fixed) space of secrets X. Both priors π and posterior distributions are elements of DX.

1) Fixed g, universally quantifiedπLg(∀,C):

Definition 4: Multiplicative g-capacity For channel C and gain functiong, the multiplicativeg-capacity is given by Lg(∀,C) = sup

π:DX

Lg(π,C) = sup

π:DX

lg(Vg[π,C]/Vg[π]). 2) Universally quantified g, fixedπ L(π,C):

Definition 5: Multiplicative π-capacity For channel C and prior π, the multiplicativeπ-capacity is given by

L(π,C) = sup

g:G1X

Lg(π,C) = sup

g:G1X

lg(Vg[π,C]/Vg[π]). It is worth pointing out that quantifying over gain functions in G1X is exactly equivalent here (in the multiplicative case) to quantifying over gain functions with range [0,1] (as used in previous work). First, gain functions with range [0,1] are in G1X. Second, any g:G1Xcan be scaled to have range [0,1], with no effect on the multiplicative leakage.

3) Universally quantified g and πL(∀,C):

Definition 6: Multiplicative capacity For channelC the general multiplicative capacity is given by

L(∀,C) = sup

g:G1X

Lg(∀,C) = sup

π:DX

L(π,C).

B. Additive capacities defined

1) Fixed g, universally quantifiedπL+g(∀,C):

Definition 7: Additive g-capacity For channel C and gain functiong, the additiveg-capacity is given by

L+g(∀,C) = sup

π:DX

L+g(π,C) = sup

π:DX

Vg[π,C]−Vg[π]. This generalises Shannon capacity, but with Shannon en- tropy replaced by some g-vulnerability.

2) Universally quantified g, fixedπ L+(π,C):

Definition 8: Additive π-capacity For channel C and prior π, the additiveπ-capacity is given by

L+(π,C) = sup

g:G1X

L+g(π,C) = sup

g:G1X

Vg[π,C]−Vg[π]. As mentioned above, restricting gain functions toG1Xdoes make a difference in additive leakage. But note that a gain function with arbitrary range can always be expressed as a gain function in G1X followed by some shifting (which has no effect on the additive leakage) and scaling (which multiplies the additive leakage by a constantk). Hence we can, if desired, simply scale L+(π,C) to account for a larger class of gain functions.

3) Universally quantified g and πL+(∀,C):

Definition 9: Additive capacity For channelC the gen- eral additive capacity is given by

L+(∀,C) = sup

g:G1X

L+g(∀,C) = sup

π:DX

L+(π,C).

C. A capacity scenario related to Dalenius’s Desideratum In addition to the six capacity scenarios considered above, we can also identify a seventh scenario that is considerably more “open ended”.

In the context of differential privacy [25], [33], a major motivation is the possibility that the adversary knows inter- esting correlations among secrets. Recall Dwork’s argument for the impossibility ofDalenius’s Desideratum—she imagines the prior knowledge “Alan Turing is two inches taller than the average Lithuanian woman.” With such knowledge, revealing information about Lithuanian women has the surprising effect of revealing information about Alan Turing.

And indeed we can consider such scenarios in the context ofg-leakage. Suppose we have a secretXwith priorπ, and an adversary interestedonlyin learning aboutX, as measured by a gain functiong:W × X →R. Now imagine a channelCfrom Y toZ, apparently having nothing at all to do withX. But still the adversary might know some interesting correlation between X andY, expressed as a joint distributionΠonX × Y(which must induce marginal distribution π on X). In this case, we can quantify how muchC(surprisingly) teaches the adversary about X. For we can extendC to a channelC from (X,Y) to Z, defined byC(x,y),z =Cy,z. (Note thatCignores the value of X.) And we can extend gain functiong to a gain function g for (X,Y), defined by g(w,(x,y)) = g(w,x); this reflects the fact that the adversary cares only aboutX. With this done, we find thatLg(Π,C) tells how much channelC leaks aboutX, given knowledge of correlations Π.

Note here that Vg[Π] = Vg[π], but Vg[Π,C] could be much larger. This means that the adversary’s knowledge of the correlations Π between X and Y does not immediately teach him anything new about X. It is only observing the output of Cthat gives him new information about X—and this in spite of the fact that channelCcompletely ignores the value ofX.

Hence it would be especially significant to establish upper bounds on the capacity (whether additive or multiplicative) on Cin terms of the capacity ofC—such bounds would give us a way to limit the harm that channelC could cause to X, no matter what correlations might subsequently be discovered to exist betweenXandY. We give such bounds in Corollaries 22 and 23 in§IX below.

VI. Calculating multiplicative capacities

We now consider the efficiency of calculating capacities in our three multiplicative scenarios.

(8)

A. Fixed g, universally quantified πLg(∀,C)

This scenario appears quite challenging. For the identity gain function gid (i.e. min-entropy leakage), we know from [20], [29] that a uniform prior always realizes the capacity.

But this is not true in general. In [11,§V-D] are a channel and gain function whose multiplicative leakage is not maximized by a uniform prior but instead by π=(1/2,1/2,0). It is unclear whether there is an efficient algorithm in this case.

B. Universally quantified g, fixed πL(π,C) Here we have discovered a complete solution:

Theorem 10: For any full-supportπ, there exists a gain functiongπ such that for anyC we haveLgπ(π,C)=ML(C).

(As a consequence,Lgπ(π,C)=L(π,C) since, by the Miracle Theorem of [11], we have ML(C)=L(∀,C).)

Proof: Letgπ:X×X →[0,1] be given by gπ(x,x) =

( a/πx if x=x

0 otherwise, (7)

wherea=minxπx. Note thatgπ is independent ofC; it can be written as a diagonal matrix whose entries are the reciprocals of the entries of π, normalized to be in [0,1]. Intuitively, gπ cancels out the “hints” coming from a skewed prior, making every element ofXequally good a priori.

Now straightforward calculations show thatVgπ[π]=aand Vgπ[π,C]=aP

ymaxxCx,y, givingLgπ(π,C)=lgP

ymaxxCx,y. The latter is the logarithm of the sum of the column maximums of C, which isML(C) by the theorem of [20], [29].

As an example, consider the imperfect cancer test from [30], with channel matrix

C positive negative cancer 0.90 0.10 no cancer 0.07 0.93

For priorπ=(1/125,124/125) the min-entropy leakage is 0, since the best guess is “no cancer”, regardless of the test result. But with gain function

gπ cancer no cancer

cancer 1 0

no cancer 0 1/124

.

we find that the gπ-leakage is log 1.83=ML(C).

Finally, note that ifπis not full support, we can just delete the rows ofCcorresponding to values ofXnot in the support.

C. Universally quantified g and πL(∀,C)

Here we also have a complete solution. By the Miracle Theorem, multiplicative leakage is maximized by gid and the uniform π, which give the min-capacityL(∀,C). In fact, by

§VI-B we can let πbe any full-support prior and letg be the

“reciprocals of π” gain functiongπ.

VII. Calculating additive capacities

We now consider the efficiency of calculating capacities in our three additive scenarios.

A. Fixed g, universally quantified πL+g(∀,C)

Here we know a number of interesting things. For gid

(which gives additive min-entropy leakage), it is shown in [29] that the additive leakage is maximized by some “corner point” distribution, i.e. one that is uniform on somesubset of X. If|X|=n, there are 2n−1 corner-point distributions, making the maximum additive leakage computable. But, because the number of corner-point distributions is exponential in the size of C, this does not give an efficient algorithm. And the following theorem shows that an efficient algorithm probably does not exist: it is NP-complete to decide whetherC’s additive gid-capacity exceeds a given threshold.10

Theorem 11: Given a channel matrixCand a threshold t, it is NP-complete to decide whetherL+g

id(∀,C)t.

Proof: Before beginning the proof, we establish some basic properties and notation for corner-point distributions.

Given a non-empty subsetSofX, the corner-point distribution πSgives probability |S|1 to the values inSand (therefore) gives probability 0 to the values inX−S. Now observe that

L+g

idS,C)

= VS,C]V[πS] “gid gives ordinary vulnerability”

= P

y:Ymaxx:X πSxCx,y−maxx:XπSx

= P

y:Ymaxx:S 1

|S|Cx,y|S|1

= |S|1(P

y:Ymaxx:SCx,y−1). Notice that P

y:Ymaxx:SCx,y is simply thesum of the column maximums of C, restricted to the rows in S. To simplify notation we let σS abbreviate this sum, allowing us to write

L+g

idS,C) = σS−1

|S| . (8)

Now we proceed with the proof. By the result in [29], it is easy to see that the decision problem “L+g

id(∀,C)t?” is in NP. For we can simply guess a non-empty subsetSand verify that L+g

idS,C)t.

We complete the NP-completeness proof by showing a reduction from the Set-Packingproblem [SP3] of [34], which is the problem of deciding whether a collection of sets U = {U1, . . . ,Un}containskpairwise disjoint sets.

Suppose that the elements of the Ui are drawn from universe {a1, . . . ,am}. We begin our reduction by constructing an n×mchannel matrix11 CU where

Ci,Uj = ( 1

|Ui| ifajUi 0 otherwise.

Now notice that ifU containskpairwise disjoint sets, k≥1, thenCUcontainskpairwise non-overlapping rows. This means that CU is a perfect channel on the corresponding k inputs, since any possible output from those inputs will uniquely deter- mine the input. Hence the additive leakage on the corner-point

10It is worth noting that this theorem is fundamentally different from the intractability results of [18]—those results assume that the channel is represented as aprogram, which inevitably leads to intractability of analysis.

In contrast, our theorem is based on the channel matrix, which gives the channel’s behaviour explicitly.

11Note that if∅ ∈ U, we do not actually get a channel matrix, sinceleads to an all-zero row. But in this case notice thatUcontainskpairwise disjoint sets iffU−{∅}containsk−1 pairwise disjoint sets.

(9)

distribution on those kinputs is k−1k , since the vulnerability is increased from 1k to 1.

We can see this more formally from equation (8) above.

For if we letSbe a set consisting of the indices ofkpairwise disjoint sets in U, then we have

L+g

idS,CU) = σS−1

k

where σS is the sum of the column maximums of CU, restricted to the rows inS. But, since the rows inS are non- overlapping, we know that each nonzero entry in each of these rows is the onlynonzero entry in its column (restricted to the rows in S), and hence is included in the sum σS. So, since each row sums to 1, we get thatσS=k, which gives

L+g

idS,CU) = k−1

k .

So far we have shown that if U contains k pairwise disjoint sets, then we have L+g

id(∀,CU) ≥ k−1k . But does the converse hold? Not necessarily, as shown by the following counterexample:

CU =











1 0 0 0 0 0 0

1 4

1 4

1 4

1

4 0 0 0

1

4 0 0 0 14 14 14











 .

In this case,Udoesnotcontain 2 disjoint sets, but nevertheless L+g

id(∀,CU)≥ 2−12 = 12, since the corner-point distribution on all 3 rows gives additive leakage of 13(52 −1)= 12.

To correct this difficulty, we need to convert U into a channel matrix in a slightly more complicated way, one which boosts the “penalty” for including an overlapping row in a corner-point distribution.

To this end, let p≥1 be the maximum cardinality of any of the sets in U.12 LetCU be embedded into a block matrix CU withp new rows and pnew columns:

CU = Ip×p 0p×m

0n×p CU

(Here I denotes the identity matrix, and 0 a zero matrix.) Extending our previous calculation, note that ifUcontains k pairwise disjoint sets, then we have

L+g

id(∀,CU) ≥ p+k−1

p+k , since the p new rows are all non-overlapping.

But we can now argue that the converse also holds. For if L+g

id(∀,CU)≥ p+k−1p+k , then there exists a set S of indices such thatL+g

idS,CU)≥ p+k−1p+k . Assume further that we chooseSof minimum sizethat can achieve this additive-leakage threshold.

Then we can argue by contradiction that the rows in S must be nonoverlapping, because if Scontains a row that overlaps some other row ofS, then we candeleteit without decreasing the additive leakage.

12In fact, the Set-Packing problem remains NP-complete even if all sets in Uhave cardinality at most 3 [34, p. 221].

For suppose that |S|= sfor s≥2, and that S is formed by deleting a row with some overlap with the other rows of S. Recall that

L+g

idS,CU) = σS−1

s ,

where σS is the sum of the column maximums of CU, restricted to the rows in S. Now observe that when we pass fromStoS, we lose the contribution toσS from the deleted row; this loss is at most p−1p , since the deleted row must have overlap of at least 1p with the other rows ofS. Hence we have

L+g

idS,CU)

= (σS−1)/(s−1)

≥ (σSp−1p −1)/(s−1)

≥ (σSp+k−1p+k −1)/(s−1) “k≥0”

≥ (σSσSs−1−1)/(s−1) “L+g

idS,CU)≥p+k−1p+k

= (S−1)(s−1)s )/(s−1)

= (σS−1)/s

= L+gidS,CU).

Having proved that the rows in Sare nonoverlapping, we now note that S must contain at least p+k rows, since a nonoverlapping set of sizeugives additive leakage u−1u , which is less than p+k−1p+k unless up+k. And, finally, at least k of the nonoverlapping rows must come from CU, which implies that U containsk pairwise disjoint sets.

We have thus shown thatUcontainskpairwise disjoint sets iffL+g

id(∀,CU)≥ p+k−1p+k . Hence indeed Set Packing polynomial- time reduces to Maximum Additive Leakage.

Given the result for gid specifically, we cannot of course expect the problem to be efficiently solvable forg in general.

But we observe that we can use the linear-programming-based algorithm for Problem 6.1 of [11] to find a maximizingπ: given channels C1 andC2, that algorithm finds a π that maximizes Vg[π,C1]−Vg[π,C2]. (It is aimed at testing whether C1’s g- leakage can ever exceedC2’s.) Here we can just letC1beCand letC2 be O, where Ois a channel with just one column,13 so that it satisfies noninterference, givingVg[π,O]=Vg[π]. Hence we get Vg[π,C1]−Vg[π,C2]=Vg[π,C]−Vg[π]=L+g(π,C).

To use linear programming, the algorithm in [11] applies a strategy-based formulation of posterior g-vulnerability. A strategyS reifiesthe choice of best guess as a channel fromY (the set of outputs ofC) toW(the set of allowable guesses).

We see then that we can formulate posterior vulnerability as Vg[π,C] = max

S

X

x,y,w

πxCx,ySy,wg(w,x) .

But notice that if we then try to compute the maximization maxπ (Vg[π,C1]−Vg[π,C2]),

we havenested maxesand aquadraticobjective function, since the entries of both πandS are variables.

However, we can assume without loss of generality that the strategy S is deterministic, which allows the algorithm

13We writeObecause it is a right-zero of channel cascading.

Références

Documents relatifs

We can confidently state that oral rehydration therapy can be safely and successfully used in treat- ing acute diarrhoea due to all etiologies, in all age groups, in all

Cherney, “Fundamental and Low Frequency Harmonic Components of Leakage Current as a Diagnostic Tool to Study Aging of RTV and HTV Silicone Rubber in Salt-Fog”,

Furthermore, as shown by Exam- ple 1, in order to determine this differential Hilbert function it may be necessary to compute derivatives of the original equations up to order ν,

As we will see in Section III, gain functions let us model a wide variety of scenarios, including those where the adversary benefits from guessing a value close to the secret,

In fact, he showed an example of systems with the same mutual information, the same a priori uncertainty, and for which the probabilities of guessing the right input a posteriori

The rewriting process is driven by the program graph which describes the source program structure: the goal is to use the structure to randomize the order according to

Thus, hedefi ned the multiplication of two cardinal numbers 1, and, although the extension of this definition to a transfinite number of cardinal numbers

The workflow model is here extended by considering tasks to be composed of steps, each of which has an execution time, leakage value, and security level.. Each one of these steps