• Aucun résultat trouvé

Learning Conditional Preference Networks with Queries

N/A
N/A
Protected

Academic year: 2021

Partager "Learning Conditional Preference Networks with Queries"

Copied!
7
0
0

Texte intégral

(1)

HAL Id: hal-00944391

https://hal.archives-ouvertes.fr/hal-00944391

Submitted on 12 Feb 2014

HAL

is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire

HAL, est

destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Learning Conditional Preference Networks with Queries

Frédéric Koriche, Bruno Zanuttini

To cite this version:

Frédéric Koriche, Bruno Zanuttini. Learning Conditional Preference Networks with Queries. Proc.

21th International Joint Conference Artificial Intelligence (IJCAI2009), Jul 2009, United States. 6 p.

�hal-00944391�

(2)

Learning Conditional Preference Networks with Queries

Frédéric Koriche LIRMM, CNRS UMR 5506 Université Montpellier II, France

frederic.koriche@lirmm.fr

Bruno Zanuttini GREYC, CNRS UMR 6072

Université de Caen Basse-Normandie, France bruno.zanuttini@info.unicaen.fr

Abstract

We investigate the problem of eliciting CP-nets in the well-known model of exact learning with equiv- alence and membership queries. The goal is to identify a preference ordering with a binary-valued CP-net by guiding the user through a sequence of queries. Each example is a dominance test on some pair of outcomes. In this setting, we show that acyclic CP-nets are not learnable with equivalence queries alone, while they are learnable with the help of membership queries if the supplied examples are restricted to swaps. A similar property holds for tree CP-nets with arbitrary examples. In fact, membership queries allow us to provideattribute- efficientalgorithms for which the query complex- ity is only logarithmic in the number of attributes.

Such results highlight the utility of this model for eliciting CP-nets in large multi-attribute domains.

1 Introduction

The spectrum of AI applications that resort on the abil- ity to reason about preferences is extremely wide, ranging from configuration softwares and recommender systems to autonomous agents and group decision-making. As many, if not most, of these applications are defined over large, multiat- tribute domains, a key challenge in preference research is to develop representation languages and elicitation techniques that cope with the exponential size of the outcome space.

Conditional preference networks(CP-nets), have emerged as an expressive language capable of representing ordinal preferences relations in a compact and structured manner [Boutilieret al., 2004]. Briefly, a CP-net is a graph where each node is labelled with a table describing the user’s prefer- ence over alternative values of this node given different values of the parent nodes. For example, the entryJb∧Pb:Sr≻Sb

might state that, all other things being equal, I prefer a red shirt than a black one if the color for both the jacket and the pants is black. The semantics of a CP-net is defined by a dominance ordering on the outcome space, derived from such reading of entries in the tables. Based on this relation, a key reasoning task isdominance testing: given a CP-netNand a pair of outcomes(o, o), determine whetherodominateso, according to the dominance ordering induced byN.

Ideally, in preference elicitation with CP-nets, the decision-maker should simply “fill the tables” by asking the user how her preference over the values of one node depends on the values of its parents. Yet, in practice, eliciting pref- erences is far from easy because the dependency graph is generally not known in advance: the decision-maker must therefore seek the interdependencies between attributes and identify a minimal set of parents for each target node. The problem is exacerbated still further by the fact that real-world applications typically involve many irrelevant attributes. For instance, it is not uncommon in recommender systems to describe products using thousands of variables, with the expectation that only a small fraction of these are crucial for describing preferences [Basilico and Hofmann, 2004;

Ziegler et al., 2008]. The decision-maker is thus required to select, within a large collection of attributes, a relatively small subset over which the network will be specified.

Such considerations bring into sharp focus the need for query learning algorithmsthat aim at extracting CP-nets by guiding the user through an appropriate sequence of queries.

A widely adopted framework for studying this issue is the model of exact learning with equivalence and membership queries [Angluin, 1988]. In essence, equivalence queries sim- ulate a form ofpassivelearning in which the decision-maker observes the user’s behavior until she finds a counterexam- ple to her hypothesis. By contrast, membership queries cap- ture a form ofactivelearning by allowing the decision-maker to ask about examples of her own choice. The utility of this model lies in the fact that rich concept classes, including Horn theories, decision trees, and some description logics, have been shown to be learnable with both equivalence queries and membership queries, while in weaker versions one can prove superpolynomial lower bounds [Angluin et al., 1992;

Bshouty, 1995; Frazier and Pitt, 1996].

In the learning model suggested by this study, the target concept is a dominance ordering on the outcome space. Each example is a preference situation involving some pair of out- comes. For amembership query, the learner supplies an ex- ample(o, o)and is told whetherodominateso, or not. For anequivalence query, the learner presents a CP-netN, and either is told thatNcorrectly identifies the target concept, or it is given a counterexample(o, o). The goal is to identify the target concept using as few resources as possible, where resources refer both to the run time and the number of queries.

(3)

From a practical perspective, one must take into account the fact that outcomes are typicallynotcomparable with an equivalent cost. As observed in [Green and Srinivasan, 1978], users can meaningfully compare outcomes if they differ only on very few attributes. Similarly, for the learner, this task can be arduous because dominance testing is generally NP-hard, even for acyclic CP-nets. Thus, our learnability results are defined in terms of aconcept classin which the target concept is chosen, and aninstance classthat circumscribes the set of examples used by equivalence and membership queries.

The key message to be gleaned from this paper is that active learning is required to correctly and efficiently ex- tract preference networks in binary-valued domains. On the one hand, acyclic CP-nets arenotlearnable with equivalence queries alone, while on the other, they are learnable with equivalence and membership queries, provided that the in- stance class is restricted to simple outcome pairs for which dominance testing takes linear time. Interestingly, a simi- lar property holds for tree-structured CP-nets by extending the instance class to arbitrary examples. When membership queries are available, we provideattribute-efficient learning algorithms for which the query complexity is linear in the size of the minimal CP-net that identifies the target concept, andlogarithmicin the total number of attributes. Such en- couraging results pave the way for fast elicitation techniques capable of extracting “small” CP-nets in “large” domains.

2 Conditional Preference Networks

The learning problems under consideration in this study are defined over a set of Boolean variablesX ={x1,· · ·, xn}. As usual, we refer toxi andxias literals. Given a literalp, we denote bypthe opposite ofp; for example, ifpiisxi, then piisxi. A termtis a conjunction of literals. Byvar(t)we denote the set of variables occurring int. A termtismaximal for a subset of variablesY ⊆Xifvar(t) =Y.

Aconditional preference rule(CP-rule) on a variablexis an expression of the formt:p≻p, wherepis a literal ofx andtis a term such thatx6∈var(t). Such a rule captures the statement “given thattholds, the valuepis preferred to the valuepfor the variablex, all other things being equal”.

Aconditional preference table(CP-table) on a variablex with respect to a setY ⊆X\ {x}is a set of CP-rules onx that associatesat mostone rulet : p ≻ pto each maximal term t forY. The CP-table is completeif exactlyone rule t:p≻pis associated to each maximal termtforY.

Aconditional preference net(CP-net) is a labelled digraph Nover a subsetvar(N)ofX, where each nodex∈var(N) is annotated with a CP-tablecpt(x)onxwith respect to the setPa(x)of parents ofxin the graph. The CP-net iscomplete if any nodex ∈ var(N)is annotated with a complete CP- table. A CP-net isacyclicif its digraph is acyclic, andtree- structuredif its digraph forms a forest, that is, a set of trees.

These notions are illustrated in the left part of Figure 1, where an acyclic complete CP-net is specified for the popular evening dress scenario. I unconditionally prefer black (, p) to white (, p) as the color of both the jacket and the pants, while my preference for a red shirt (s) versus a white one (s) depends on the combination of jacket and pants.

≻ p≻p

?>=<

89:;

%%J

JJ JJ

J ?>=<89:;p

yyssssss

?>=<

89:;s

∧p s≻s

∧p s≻s

∧p s≻s

∧p s≻s

ps

ps

((

vv

ps

ps

ps

((

ps

vv

ps

ps

Figure 1:A CP-net and its preference graph for “evening dress”

Anoutcomeis a maximal termoforX. Given a termt, we writeo[t]for the outcome obtained by makingoagree with t, i.e. o[t] = {p : p ∈ torp ∈ o−t}. For example, if o=x1x2x3andt=x1x2, theno[t] =x1x2x3. An outcome osatisfiesa termtifo =o[t]. We write0(resp. 1) for the outcome that assigns0(resp.1) to every variable inX.

The ceteris paribus semantics of a CP-rule t : p ≻ pon a variablex can be described as follows: if an outcome o satisfies the condition tand assigns the valueptox, then it is preferred to an outcomeowhich differ fromoonly in that is assigns the valueptox. In formal terms, we say thatois preferredtoofort:p≻pifo=o[t∧p]ando =o[p]. In this case,(o, o)is called amodelof the rulet:p≻p.

Given a CP-netN and two outcomesoando, we say that odominatesoforNif there is a sequence(o1,· · ·, om)such thato1=o,om = oand for eachi: 1≤i < m,(oi+1, oi) is a model of some CP-rule ofN. In this case,(o, o)is called amodelofN, and(o1,· · ·, om)animproving sequencefrom o too. Thedominance orderingofN, denoted≻N, is the set of all models ofN. For example, the dominance ordering of the CP-net for evening dress is the transitive closure of the digraph displayed in the right part of Figure 1.

A CP-net N isconsistentif there is no outcomeowhich dominates itself, i.e.o≻N o. IfNis consistent, then≻N is a strict partial order over the outcome space. As reported in [Boutilieret al., 2004], any acyclic CP-net is consistent.

Given two CP-netsNandN, we say thatNsubsumesN if for any CP-rulet:p≻pinN, there is a CP-rulet:p≻p inN such thatt ⊆t. A CP-netN isminimalif there is no distinct netNsubsumed byNand for which≻N=≻N. For example, the CP-net in Figure 1 is minimal.

Finally, thesize |r| of a CP-ruler will be the number of occurrences of literals in its definition. Specifically, ifris of the formt:p≻p, then|r|=|t|+2. Thesize|N|of a CP-net N will be the sum of|r|over all rulesrinN. For example, the size of the CP-net in Figure 1 is2 + 2 + 4×(2 + 2) = 20.

3 Exact Learning with Queries

The learning criterion expected in this study is that ofexact identification, which is achieved if the learner can infer a CP- net that correctly represents the target concept. Aconceptis a strict partial ordering≻on the outcome space. Arepresenta- tion classis a setN of CP-nets. A concept≻isrepresentable byN if there is a CP-netN ∈ N such that≻N=≻. The

(4)

concept classCN overN is the set of all concepts that are representable byN. Thedescription sizeof a concept≻in CN will be the minimum of|N|over all representationsNof

≻inN. Finally, an instance orexampleis a pair(o, o)of outcomes, and aninstance classis a setOof examples.

Let≻be a target concept of some concept classCN, and Obe an instance class. The learner may extract information about ≻using two types of queries. A membership query MQoverOtakes an example(o, o) ∈ O and returnsYes if o ≻ o, andNo otherwise. Anequivalence query EQ overO takes a CP-netN ∈ N, and returnsYes ifN is a representation of≻, or returns a counterexample(o, o)∈ O otherwise. The counterexample(o, o)ispositiveifo ≻ o ando⊁N o, andnegativeifo⊁oando≻N o.

Definition 1.An algorithmAis aquery learning algorithm for a concept classCN with respect to an instance classO if there are two polynomialspandq such that, for any tar- get concept≻inCN, afterp(n)membership and equivalence queries overO, and total running time inq(n),Aoutputs a representationN ∈ N of ≻. It is attribute-efficient if the number of queriesp(n)is polynomial in the description size of≻, but only polylogarithmic in the number of variablesn.

Given an instance class O, we say that a concept class CN isattribute-efficiently learnable if there is an attribute- efficient query learning algorithm forCN with respect toO.

Clearly, the strength of query-directed learning lies in membership queries, which model not only the interaction with a user, but also the careful crafting of experiments by a decision-maker in order to observe the response of the user.

In order to demonstrate that a class of CP-nets isnotlearnable with equivalence queries alone, we shall use the technique of approximate fingerprintsintroduced in [Angluin, 1990].

Intuitively, a concept class CN has approximate finger- prints if it includes a setCN such that for each hypothesis NinCNsupplied by the learner, the user can choose a coun- terexample forN that eliminates only a superpolynomially small fraction of candidate concepts inCN. By repeating this process, the learner cannot be certain of the target concept in CN after only polynomially many equivalence queries.

A pair(o, o)of outcomes is called an α-fingerprintof a concept classCN according to some concept≻1if

|{≻2∈ CN :o≻2oiffo≻1o}|

|CN| < α

Formally, a concept classCN hasapproximate fingerprints with respect to an instance classOif for any polynomialp(n), CNincludes a subclassCN such that for any sufficiently large n,CN contains at least two concepts, and for all concepts≻in CN of description size bounded byp(n), there is an example inOwhich is a p(n)1 -fingerprint ofCN according to≻.

4 Learning Acyclic CP-nets with Queries

Acyclic CP-nets take a central part in preference research by providing the right level of expressivity for many real-world applications, while remaining tractable for certain reasoning tasks such as outcome optimization [Domshlaket al., 2001;

Boutilieret al., 2004]. We begin with some useful properties.

Two outcomes form a swapif they differ in the value of only one variable. Such examples correspond to simple sit- uations of the form “for this car I prefer it red than white”, where the color is one of the multiple attributes of the car.

The next property states that, in acyclic CP-nets, swaps can be retrieved in linear time by simple rule matching.

Lemma 1.LetNbe an acyclic CP-net and(o, o)be a swap.

Theno ≻N o if and only if there is a CP-rulerinN such that(o, o)is a model ofr.

Proof. Theif direction is immediate. Conversely, ifo≻N o, then there is an improving sequence fromotooinN. As- sume w.l.o.g. thatosatisfiesx1ando=o[x1]. Using the suf- fix fixing rule [Boutilieret al., 2004, Section 5.1], we can also assume that the sequence affects onlyx1 and its ascendants inN. By acyclicity, one of those ascendantsxk is modified but has none of its parents modified in the sequence. Thus, xk cannot be modified back to its value ino, which entails xk =x1. So onlyx1is modified, which concludes.

A CP-netN is aminimal representationof a concept≻if Nrepresents≻and|N|is minimal.

Lemma 2.For the classCACYof acyclic CP-nets, the minimal representation of any concept is unique.

Proof. Let ≻be a concept inCACY andN a representation of≻satisfying the following: for any variablesxandy, and any ruler ∈ N onx, ify is a parent ofxinN, then there is a model(o, o)of r such that exactly one of(o[y], o[y]) and(o[y], o[y]) is not a model ofr. This amounts to say thatyis arelevantparent ofx. Clearly, such a representation exists: take any representation and remove all irrelevant par- ents. Now letN be any representation of≻; we show that NsubsumesN. Letr∈N be a rule of the formt :p≻p.

Any swap(o, o)for whicho = o[t∧p]ando = o[p]is a model ofr. Since≻N=≻N, by Lemma 1,(o, o)must be a model of some rule r inN of the formt : p ≻ p. If t*t, then there is a variabley ∈var(t)\var(t)such that (o[y], o[y])and(o[y], o[y])are both models ofr, but exactly one of these pairs is not a model ofr, contradicting the fact that≻N=≻N. Sot⊆t, and hence,NsubsumesN.

We now show that acyclic and complete CP-nets have approximate fingerprints, even if examples are restricted to swaps. Thus, by applying Theorem 1 in [Angluin, 1990], they arenotlearnable with equivalence queries alone.

Theorem 1.The classCACC of acyclic complete CP-nets has approximate fingerprints with respect to swap examples.

Proof. LetCACC be the class of all concepts represented by a CP-netN withlognroot nodesxj pointing to the same fixed child nodex1. Each tablecpt(xj)has the rulexj ≻xj. The tablecpt(x1)includes one rules:x1 ≻x1, wheresis the conjunction of all positive literals inPa(x1), andn−1 ruless :x1 ≻x1, wheresis any maximal term ofPa(x1) with at least one negative literal. ClearlyN is acyclic and complete. Furthermore,|CACC|= logn−1n

. Now, letp(n) =nk for some constantk, and letN ∈ CACCwith|N| ≤p(n).

(5)

The fingerprint(o, o)is defined as follows. IfN does not include any rule of the formt :x1 ≻x1, then we leto =1 ando =o[x1]. Theno⊁N o buto≻ofor any concept≻ inCACC . So(o, o)is anα-fingerprint ofCACC for anyα >0.

Now, ifN includes a rule of the formt:x1≻x1, thenois any outcome satisfyingt∧x1and containingklognpositive literals excluding x1, which can be constructed as follows.

BecauseN is complete, the size of its table on x1 is more than2|t|. Since|N| ≤ nkwe get|t| ≤klogn. Thus,ocan be constructed by satisfyingtfirst and filling the rest as nec- essary. Again,o = o[x1]. Soo ≻N o ando hasklogn positive literals (excludingx1). Hence, the number of con- cepts≻inCACCfor whicho≻ois kloglognn

. Usinga−ib−iab,

|{≻∈ CACC :o≻o}|

|CACC | =

logn−1

Y

i=0

klog(n−i)

n−1−i ≤(klogn)logn (n−1)logn Taking logarithms, this proportion is less than n1k if and only if logn−1n > k2k, which is true for sufficiently largen.

When membership queries are available, we can provide an attribute-efficient algorithm for learning acyclic, and pos- siblyincomplete, CP-nets, provided that the supplied exam- ples are restricted to swaps. As specified in Algorithm 1, the learner iteratively updates her hypothesisNby asking equiva- lence queries. On seeing a counterexample(o, o), the learner checks whetherN includes a rule that covers either(o, o)or (o, o). If this is indeed the case, she asks membership queries in order to refine the condition of that rule. Otherwise, she ex- pands her net with a new rule. Here, each rulerof the form t:pi ≻piis associated with an outcomeor, called thesup- portofr, and such that(or[pi], or[pi])is a model ofr.

The key routine SEARCHPARENT finds a new parent of some misclassifying ruler, using only a logarithmic number of membership queries. Using the supportorofrand the last counterexample(o, o), it operates a binary search on the se- quence(o1,· · ·, on)whereojis formed by the firstjliterals occurring inorand the lastn−jliterals occurring ino. As shown in the lemma below, SEARCHPARENTis guaranteed to find a new parent in the ruler, by maintaining the invari- ant that for each explored subsequence(oa,· · ·, ob), we have bothoa[pi]⊁oa[pi]andob[pi]≻ob[pi]in the target concept.

Lemma 3.Let≻ be a concept ofCACY, oa, ob be two out- comes, andpi, pibe a pair of opposite literals for some vari- ablexi. If we haveoa[pi]⊁oa[pi]andob[pi]≻ob[pi], then there is a parentxjofxiin the minimal representationNof

≻whose value is different inoaandob.

Proof. Consider the outcome sequence (o0,· · ·, on)where oj = oa[tj]withtj the conjunction of the firstj literals in ob. Sinceo0=oaandon=ob, there is somej >0such that oj−1[pi] ⊁ oj−1[pi]andoj[pi] ≻ oj[pi]. So there is a rule t : pi ≻piinNsuch thatoj satisfiestbutoj−1does not.

Since they differ only onxj, we getxj∈var(t).

Lemma 4.Let≻be a target concept in the classCACY. Then Algorithm 1 maintains an acyclic CP-netNwhich is always subsumed by the minimal representationNof≻.

Algorithm 1: Learning Acyclic CP-nets N ←∅

while(o, o)←[EQ(N)6=Yes]do

letxibe the variable s.t.pi∈oandpi∈o if(o, o)is a model of some rulehr, oriinN then

(o, o)←(o, o)

if(o, o)is a model of some rulehr, oriinN then /The counterexample is negative/

xj←SEARCHPARENT(xi, pi, o, or,0, n) foreach rulehr, oriin the table ofxiexpand the condition ofrwith the literal ofxjinor

else

/The counterexample is positive/

add the ruleht:pi≻pi, oitoNwheretis the projection ofoonto the setPa(xi)inN returnN

Procedure SEARCHPARENT(xi, pi, o, o, a, b) ifa=b+ 1then returnxb

j ← ⌊(a+b)/2⌋

oj←o[tj]wheretjare the firstjliterals occurring ino ifMQ(oj[pi], oj[pi]) =Yesthen

return SEARCHPARENT(xi, pi, o, o, a, j) else

return SEARCHPARENT(xi, pi, o, o, j, b)

Proof. Initially, N = ∅, so the property holds. Now, con- sider an iteration of the main loop and suppose by induction hypothesis (I.H.) thatN is subsumed byN before calling EQ. If EQ returns a positive counterexample(o, o), then by Lemma 1 N includes a rule t : pi ≻ pi for which o = o[t∧pi] ando = o[pi]. ThenN is expanded with t: pi ≻piwheretis the projection ofoonto the parent set Pa(xi)ofxiinN. Since (by I.H.)Nis subsumed byN, we havePa(xi)⊆Pa(xi), wherePais the parent set inN. Thereforet⊆t, and hence, the invariant is preserved.

Dually, if EQ returns a negative counterexample(o, o), then by Lemma 1Nincludes a rulerof the formt:pi≻pi

for whicho=o[t∧pi]ando =o[pi]. By construction, for the supportorofrwe also haveor=or[t∧pi], andor≻or[pi].

So by Lemma 3, SEARCHPARENTreturns a parentxjofxi

inN. Now, letrbe any rule of the formt :pi ≻pi. Since the supportor ofrsatisfiesor[pi]≻or[pi], by I.H., there is a rulet :pi≻piinNsuch thatt⊆tandor satisfies t. This together with the fact thatxjis a parent ofxiinN ensurest∪ {pj} ⊆t, so the invariant is preserved.

Now let us examine the complexity of Algorithm 1. For equivalence queries, each counterexample allows us to find a new rulet : pi ≻ pi or a new literalpj of some rule in the minimal representationNof the target concept. Because the hypothesisN is always subsumed byN, this can happen at most|N|times. For membership queries, at mostlognof these are used in each call of SEARCHPARENT, which always uncovers a new parent of some variable. So the number of these queries is at mostelogn, wheree=P

iPa(xi).

(6)

Theorem 2.The classCACY is attribute-efficiently learnable from equivalence and membership queries over swaps: any concept≻inCACYcan be identified in polynomial time, using at most|N|+ 1equivalence queries ande⌈logn⌉member- ship queries, whereN is the minimal representation of≻, andeis the number of edges inN.

5 Learning Tree CP-nets with Queries

Binary-valued tree-structured CP-nets constitute a restricted, yet important, class of preference networks for which dom- inance testing on arbitrary pairs of outcomes is solvable in quadratic time using a backtrack-free search technique [Boutilieret al., 2004]. It is therefore legitimate to study the learnability issues of this class in the general setting where the examples supplied to the learner are arbitrary preference situations. In this context, we first show that tree-structured CP-nets are not learnable with equivalence queries alone.

Theorem 3.The classCTREEof tree CP-nets has approximate fingerprints with respect to arbitrary outcome pairs.

Proof. We assume w.l.o.g. thatnis even to avoid floors and ceilings. To each permutationπof(x1,· · · , xn)we associate the smallest set of rulesNπsuch thatxπ(1)≻xπ(1)is inNπ, and for eachi > 1,Nπ includesxπ(i−1) : xπ(i) ≻ xπ(i). LetCTREE be the class of all concepts represented by someNπ

specified as above. Clearly,|CTREE |=n!. Now, letp(n) =nk for some constantkand letN ∈ CTREEwith|N| ≤p(n).1

The fingerprint(o, o)is defined as follows. Assume first that there is an outcomeo1containing at leastn/2ones and such that(o1,0)is a model ofN. Then there is an improving sequence from0too1inN, and since variables are flipped one by one, it must contain an outcomeowith exactly n/2 ones. Moreover, by constructiono≻N 0holds. Leto =0. We claim that(o, o)is an n1k-fingerprint ofCTREE w.r.t. ≻N. Indeed, a conceptNπ inCTREE has(o, o)as a model if and only if the firstn/2variables according toπare exactly those assigned1byo. Otherwise, any improving sequence inNπ

should flip at least one variable assigned0by bothoando, with no way back, a contradiction. It follows that there are (n/2)!(n/2)!concepts inCTREEwith(o, o)as a model, hence

|{≻∈ CTREE:o≻o}|

|CTREE | =

n 2!2

n! = n

n 2

−1

√2n 2n which is clearly less than n1k for sufficiently largen. Now, assume that there is noo1as above. Leto =1ando =0. Soo ⊁N o, buto≻oholds for every concept≻inCTREE . Hence,(o, o)is anα-fingerprint ofCTREEfor anyα >0.

As further evidence for the utility of membership queries, we now give an attribute-efficient algorithm for eliciting tree CP-nets in presence ofarbitraryexamples. Let△(o, o)be the set of all variables whose value differ in two outcomeso ando. Algorithm 2 uses the fact that considering only vari- ables in∆(o, o)and their ascendants is enough for searching

1For a tree CP-netN,|N| ≤6|var(N)|(two rules of size3per variable), so ifp(n)≥6n,Ncan be any tree CP-net.

Algorithm 2: Learning tree CP-nets N ←∅

while(o, o)←[EQ(N)6=Yes]do

/The counterexample is necessarily positive/ for eachx∈ △(o, o)do PROPAGATE(x) returnN

Procedure PROPAGATE(x) ifx∈var(N)then return var(N)←var(N)∪ {x}

foro, p∈ {0,1} × {x, x}do prefo,p←MQ(o[p], o[p]) if pref0,p=pref1,p=Yesfor somep∈ {x, x}then

parents ←∅

else ifpref0,p=Yesfor somep∈ {x, x}then parents ← {SEARCHPARENT(x, p,1,0,0, n)} else ifpref1,p=Yesfor somep∈ {x, x}then

parents ← {SEARCHPARENT(x, p,0,1,0, n)} else

parents ←∅

add toNa table forxwith parentsparents and the rules satisfied by theprefo,p’s

ifparents ={y}then PROPAGATE(y)

an improving sequence fromotoo. Thus, on seeing a coun- terexample(o, o), the learner computes the tables for each such variable. Because any variable has at most one parent, its table can be found using few membership queries.

From a practical perspective, note that the examples used in MQ’s are restricted to swaps in order to minimize the cog- nitive effort spent by the user in comparing outcomes.2 Lemma 5.LetNbe a minimal representation of the target concept inCTREE. Then PROPAGATEis called only on vari- ablesxinvar(N), and always extracts the table ofxinN. Proof. By construction, when PROPAGATE(xi) is called by Algorithm 2, we must havex ∈ △(o, o)ando ≻ o. So x ∈ var(N), because otherwise it would have no table in Nand hence, its value could not change along any improv- ing sequence fromotoo. Now givenx∈var(N), we show that PROPAGATEcomputes the right set of parents forx.

First, let MQ(0[p],0[p]) = MQ(1[p],1[p]) =Yes. Then by Lemma 1 there is a rulet :p≻pinNsuch that both0 and1satisfyt. Hencetis empty. Now to the second case.

HereMQ(0[p],0[p]) = YesandMQ(1[p],1[p]) = No, so by Lemma 3 there is a parent yofxinN, which is found by SEARCHPARENT. The third case is symmetric. In the last case, all queries answer No, so there is no rule onxinN, and hence,xhas no parent inN.

Consequently, in all cases PROPAGATEcomputes the right set of (at most one) parents. Because each possible rule is validated by one of the queriesMQ(o[p], o[p]), the table com- puted forxis the correct one. Furthermore, since each recur- sive call of PROPAGATEis on a variableywhich is the parent of some variable invar(N), we havey∈var(N).

2In addition, the outcomes1and0used in PROPAGATEcan be replaced by any suitable pair(o, o)for whicho={p:p∈o}.

(7)

By Lemma 5, it follows that all counterexamples supplied to the learner are positive. Moreover, from the structure of the algorithm it is easily seen that after treating (o, o), the hypothesisNcontains the correct tables for all ascendants of all variables in△(o, o). This together with the suffix fixing principle [Boutilieret al., 2004, Section 5.1] ensures thatN now agrees witho≻o, and so that the algorithm converges.

Concerning the complexity of our learning algorithm, the number of equivalence queries is at most|var(N)|+ 1, be- cause each counterexample allows the learner to treat at least one new variable inN. Likewise, the routine PROPAGATE

treats each variable invar(N)exactly once, using at most logn+ 4membership queries for computing theprefo,p’s plus searching for a parent. Finally, the hypothesis main- tained by the learner is always a subtree ofN, and hence, dominance testing can be evaluated in quadratic time.

Theorem 4.The classCTREE is attribute-efficiently learnable from equivalence and membership queries over arbitrary out- come pairs: any concept≻in this class can be identified in polynomial time using at mostk+ 1equivalence queries and k⌈logn⌉+ 4kmembership queries, wherekis the number of nodes in the minimal representation of≻.

6 Discussion

Along the lines of making query-directed learning applicable to preference elicitation, we have provided a model for learn- ing preference networks from equivalence and membership queries, together with significant learnability results. Taking into account the cognitive effort required by human users to answer queries, our model is distinguished by the close way in which it integrates learning and dominance testing, and the insistence on having convergence bounds that are polynomial in the minimal description size of the target concept, but only polylogarithmic in the total number of attributes. In essence, our results reveal that membership queries are essential for extracting both acyclic CP-nets from restricted outcome pairs, and tree CP-nets from arbitrary outcome pairs. Importantly, the examples used by these queries can be limited to “swaps”

in order to facilitate their comparison.

To the best of our knowledge, this work provides the first connection betweenactivelearning and graphical preference languages. Some authors, though, have recently focused on passivelearning, where the goal is to extract a CP-net from a set of examples [Athienitou and Dimopoulos, 2007]. Yet, in this “offline” passive learning model, the problem of finding an acyclic CP-net consistent with a set of arbitrary outcome pairs is NP-hard, even if the dependence graph is known in advance [Lang and Mengin, 2009]. Note that in the “online”

passive learning model, acyclic CP-nets are not predictable with a polynomial mistake bound, even if examples are re- stricted to swaps; this is a direct corollary of Theorem 1, that follows from a standard conversion between online learning and exact learning with EQ’s alone [Littlestone, 1988].

Perhaps the closest framework to ours is due to Sachdev [2007], who investigates some preference logics in different learning models. Although encouraging, his results do not take into account the cost of dominance testing, and the query complexity grows exponentially with the size of rules.

A direction of research that naturally emerges from our study is to extend the learnability results to larger classes of preference networks. Algorithm 1 can provide a starting point for attacking multi-valued acyclic CP-nets. Similarly, Algo- rithm 2 might be extended to directed-path singly-connected CP-nets. Less obvious, however, is to efficiently learncyclic CP-nets even with swap examples. Finally, the problem of revisingCP-nets with queries looks challenging.

Acknowledgments: Thanks to the referees for helpful comments. This work was supported by the ANR project ANR-06-BLAN-083-02.

References

[Angluinet al., 1992] D. Angluin, M. Frazier, and L. Pitt.

Learning conjunctions of Horn clauses. Machine Learn- ing, 9:147–164, 1992.

[Angluin, 1988] D. Angluin. Queries and concept learning.

Machine Learning, 2(4):319–342, 1988.

[Angluin, 1990] D. Angluin. Negative results for equiva- lence queries.Machine Learning, 5:121–150, 1990.

[Athienitou and Dimopoulos, 2007] F. Athienitou and Y. Di- mopoulos. Learning CP-networks: A preliminary investi- gation. InAdv. in Preference Handling (M-PREF), 2007.

[Basilico and Hofmann, 2004] J. Basilico and T. Hofmann.

Unifying collaborative and content-based filtering. In ICML’04, pages 9–17, 2004.

[Boutilieret al., 2004] C. Boutilier, R. Brafman, C. Domsh- lak, H. Hoos, and D. Poole. CP-nets: A tool for represent- ing and reasoning with conditionalceteris paribusprefer- ence statements.J. Artif. Intell. Res., 21:135–191, 2004.

[Bshouty, 1995] N. Bshouty. Exact learning boolean func- tions via the monotone theory.Information and Computa- tion, 123(1):146–153, 1995.

[Domshlaket al., 2001] C. Domshlak, R. Brafman, and S. Shimony. Preference-based configuration of web page content. InIJCAI’01, pages 1451–1456, 2001.

[Frazier and Pitt, 1996] M. Frazier and L. Pitt. Classic learn- ing.Machine Learning, 25(2-3):151–193, 1996.

[Green and Srinivasan, 1978] P. Green and V. Srinivasan.

Conjoint analysis in consumer research: Issues and out- look. J. Consumer Research, 5(2):103–123, 1978.

[Lang and Mengin, 2009] J. Lang and J. Mengin. The com- plexity of learning separable ceteris paribus preferences.

InIJCAI’09 (these proceedings), 2009.

[Littlestone, 1988] N. Littlestone. Learning quickly when ir- relevant attributes abound: A new linear-threshold algo- rithm.Machine Learning, 2(4):285–318, 1988.

[Sachdev, 2007] M. Sachdev. On learning ofceteris paribus preference theories. Master’s thesis, Graduate Faculty of North Carolina State University, 2007.

[Ziegleret al., 2008] C. Ziegler, G. Lausen, and J. Konstan.

On exploiting classification taxonomies in recommender systems.AI Communications, 21(2-3):97–125, 2008.

Références

Documents relatifs

Since it is easy, and reasonable, to further impose that the learned CP-net has a digraph with a small degree k (see the experiments), and hence that m is in O(n k ), we finally get

This paper presented SANTA, the first algorithm for continuous top-k processing over information streams in a social network context, able to handle both mes- sage publication and

Abstract. In this article we propose a probabilistic framework in order to study the fair division of a divisible good, e.g. Our framework follows the same idea than the

In the litera- ture, there is two major approaches to handle the problem of answering skyline queries over datasets with attributes having dynamic order, (i) algorithms which, given

By contrast, acyclic CP-nets are what is called attribute-efficiently learnable when both equivalence queries and membership queries are available: we indeed provide a

 This shows that, in constructing the simple system equivalent to a given system, we can use TP-Datalog queries instead of RTPQs in the input-test and input-query nodes

Stepwise tree automata were proved to have two nice properties that yield a simple proof of the theorem. 1) N-ary queries by UTAs can be translated to n-ary queries by stepwise

The result is a private database retrieval protocol satisfying a highly desirable tradeoff of privacy and efficiency properties; most notably: (1) no unintended information is leaked