preference modelling and...

70
Languages for Preferences Utility-based Representations Preference Elicitation / Learning Preference Modelling and Learning Paolo Viappiani LIP6 - CNRS, Université Pierre et Marie Curie www-desir.lip6.fr/viappianip/ DMRS workshop, Bolzano 18 September 2014 1/67 Paolo Viappiani Preference Modelling and Learning

Upload: others

Post on 03-Oct-2020

2 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Preference Modelling and Learning

Paolo Viappiani

LIP6 - CNRS, Université Pierre et Marie Curiewww-desir.lip6.fr/∼viappianip/

DMRS workshop, Bolzano

18 September 2014

1/67Paolo Viappiani Preference Modelling and Learning

Page 2: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Outline

1 Languages for Preferences

2 Utility-based Representations

3 Preference Elicitation / LearningStandard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

2/67Paolo Viappiani Preference Modelling and Learning

Page 3: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Preference Handling Systems are Everywhere

Not only recommender systemsComputational advertisementIntelligent user interfacesCognitive assistantsPersonalized medecinePersonal Robots

What the theory has to say about preferences?3/67

Paolo Viappiani Preference Modelling and Learning

Page 4: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

What are Preferences?

Preferences are “rational” desires.Preferences are at the basis of any decision aiding activity.There are no decisions without preferences.Preferences, Values, Objectives, Desires, Utilities,Beliefs,...

4/67Paolo Viappiani Preference Modelling and Learning

Page 5: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Are We Rational Decision Makers?

NO. Human decision makers are (often) irrational andinconsistent.(work of Nobel prize winner Daniel Kahneman)Moreover preferences are often constructed during the“decision process” itself (considering specific examples)

5/67Paolo Viappiani Preference Modelling and Learning

Page 6: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Binary relations

Preference Relation�⊆ A× A is a reflexive binary relationx � y stands for x is at least as good as y� can be decomposed into an asymmetric and a symmetric part

Asymmetric part

Strict preference �: x � y ∧ ¬(y � x)

Symmetric part

Indifference ∼1: x � y ∧ y � x

Incomparability ∼2: ¬(x � y) ∧ ¬(y � x)

6/67Paolo Viappiani Preference Modelling and Learning

Page 7: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Preference Statements

People will not directly provide their preference relation �.Rather, they will provide statements about preferred states ofaffair.Consider sentences of the type:

I like red shoes.I do not like brown sugar.I prefer Obama to McCain.I do not want tea with milk.Cost is more important than safety.I prefer flying to Athens than having a suite at Istanbul.

7/67Paolo Viappiani Preference Modelling and Learning

Page 8: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Representation Problem

Often impossible to state explicitly the preference relation (�), especiallywhen A is large→ need for a compact representation!

Logical Languages

Weighted Logics

Conditional Logics

...

Graphical Languages

Conditional Preference networks (CP nets)

Conditional Preference networks with trade-offs

Generalized Additive Independence networks

8/67Paolo Viappiani Preference Modelling and Learning

Page 9: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Representation Languages

A Compact representation is useful so that preferencescan be formulated with statements that encompass severalalternativesA preference statement: “I prefer red cars over to bluecars”But....What does this exactly mean? All red cards arepreferred to blue cars? Some red cards are preferred toblue cars? There is at lest one red car preferred to a bluecar? I prefer the average red car to the average blue car?The need of a principled “semantic” for preferencestatements

9/67Paolo Viappiani Preference Modelling and Learning

Page 10: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Logical LanguagesPreference Logics

Logical languages for preferences aim at giving a “semantic” topreference statements

Φ is a preference formula (for example: color red)

In logic you write x � Φ to say that x has the “feature” expressed byformula Φ

Mod(Φ) are the alternative where Φ holds

Von Wright semantics

The statement “I prefer Φ to Ψ” actually means preferring the state of affairsΦ ∧ ¬Ψ to Ψ ∧ ¬Φ

Still not enough...

Several preference semantics: strong, optimistic, pessimistic, opportunistic,ceteris paribus

10/67Paolo Viappiani Preference Modelling and Learning

Page 11: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

From Statements to Relations: SemanticsPreference Statements

I prefer Φ to Ψ (Φ and Ψ are propositional formula)

Preference for Φ ∧ ¬Ψ over Ψ ∧ ¬Φ (von Wright’s interpretation)

Common case boolean preferences, where Φ is a variable and Ψ itsnegation (I prefer furnished apartment rather the unfurnished)

Different Semantics

Let � be a preference relation. � satisfies the preference statements if itholds x � y for:

for every x, y ∈ A : x � Φ ∧ ¬Ψ, y � Ψ ∧ ¬Φ (Strong semantics)

iff x � Φ ∧ ¬Ψ, y � Ψ ∧ ¬Φ and additionally they have the same evaluation for the other variables(Ceteris Paribus)

x and y are maximal elements of� satisfying Φ ∧ ¬Ψ and Ψ ∧ ¬Φ respectively (Optimistic)

x and y are minimal elements of� satisfying Φ ∧ ¬Ψ and Ψ ∧ ¬Φ respectively (Pessimistic)

x is maximal, y minimal elements of� satisfying Φ ∧ ¬Ψ and Ψ ∧ ¬Φ respectively (Opportunistic)

11/67Paolo Viappiani Preference Modelling and Learning

Page 12: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Ceteris Paribus: From Statements to Networks

Preferential IndependenceKey notion in preference reasoning. It is analogous toprobabilistic independence.CP preference a preferred to a′ ceteris paribus iff ab � a′b ∀bConditional Preference a preferred to a′ given c iff ab′ � a′b′c

for a given c

CP networksThe notion of conditional preferential indipendence constitutesthe main building block to develop graphical models forcompactly representing complex preferences→ CP networks

12/67Paolo Viappiani Preference Modelling and Learning

Page 13: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

CP nets

Formalization

Each variable X is associated with a set of parents Pa(X ) and a conditionalpreference table (CPT).The CPT assigns, for each combination of values of the parents, a total orderon the values that X can take.

CP-nets

Research question: is assignment xpreferred to y ? how to find undominatedassignment of variables to values?

Start by nodes with no parents, assign bestvalue, look at the children,..

Technique of “worsening flip” sequence

Notice: strong analogy with Bayesian networks

13/67Paolo Viappiani Preference Modelling and Learning

Page 14: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Utility Representation

Utility function u : X → [0,1]

Ideal item x> such that u(x>) = 1 and u(x⊥) = 0 (scaling)Representing, eliciting u difficult in explicit formFlat utility representation often unrealistic

Color Shape Position Weight ... Utilityitem1 red round top 1kg ... 0.82item2 blue square top 2kg ... 1item3 green square bottom 5kg ... 0.96

... ... ... ... ... ... ...

14/67Paolo Viappiani Preference Modelling and Learning

Page 15: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Additive Utility Functions

Additive representation (common in MAUT)Sum of local utility functions ui over attributes (or local value functions vi

multiplied by scaling weights)Exponential reduction in the number of needed parameters

u(x) =n∑

i=1

ui (xi ) =n∑

i=1

αivi (xi ) (1)

Color v1

red 1.0blue 0.7

green 0.0

Shape v2

round 0square 0.2

star 1

Importance for attribute “color”: α1 = 0.2, for “shape”: α2 = 0.3.Notice: many algorithms in the recommender system community (forexample matrix factorization techniques) implicitly assume an additive model!

15/67Paolo Viappiani Preference Modelling and Learning

Page 16: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Generalized Additive Utility

Sum of local utility functions uI over sets of attributes (or local valuefunctions vi multiplied by scaling weights)

Higher descriptive power than strictly additive utilities, while still havinga manageable number of parameters

u(x) =m∑

i=1

uJi (xJi ) =n∑

i=1

αJi vJi (xJi ) (2)

where Ji is a set of indices, xJi the projection of x on Ji and m thenumber of factors.

Color Shape vcolor,shapered round 0.9red square 1.0red star 0.5blue round 0.4... ... ...

Position vpositiontop 1

bottom 0

Importance for factor “color+shape”: αJ1 = 0.2, for “position”: αJ2 = 0.3.16/67

Paolo Viappiani Preference Modelling and Learning

Page 17: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Utility of Money

Money is a “special” attribute with monotonicpreference

Expected Monetary Value (EMV) 6= Expected Utility(note: explanation to the St. Petersburg paradox)

For most people: U(100$) > 0.5U(200$)

U(money) is concave (most common case):risk averse decision makerU(money) is linear: risk neutralU(money) is convex: risk seeker

Concept of Certainty Equivalent (CE)

Quasi-linear utility scale: utility expressed in amonetary scale

Moreover, influence of “status quo”→ Prospect Theory [Kahneman]

17/67Paolo Viappiani Preference Modelling and Learning

Page 18: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Classic Approaches for Utility Elicitation

Assessment of multi attribute utility functionsTypically long list of questionsFocus on high risk decisionGoal: learn utility parameters (weights) up to asmall error

Which queries?Local: focus on attributes in isolationGlobal: compare complete outcomes

Standard Gamble Queries (SGQ)Choose between option x0 for sure or a gamble< x>, l , x⊥,1−l >(best option x> with probability l , worst optionx⊥ with probability 1−l)

18/67Paolo Viappiani Preference Modelling and Learning

Page 19: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

(Standard) Elicitation of Additive Models

Consider an attribute (for example, color)Ask for the best value (say, red)Ask for worst value (gray)Ask local standard gable for each remaining color to assessit local utility value (value function)

Bound queries can be askedRefine intervals on local utility values

Scaling factorsDefine reference outcomeAsk global queries in order to assess the difference in utilityoccurring when, starting from the reference outcome,“moving” a particular attribute to the best / worst

19/67Paolo Viappiani Preference Modelling and Learning

Page 20: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Automated Elicitation vs Classic Elicitation

Problems with the classic view

Standard gamble queries (and similar queries) are difficult to respond

Large number of parameters to assess

Unreasonable precision required

Cognitive or computational cost may outweigh benefit

Automated Elicitation and Recommendation

Important points:

Cognitively plausible forms of interaction

Incremental elicitation until a decision is possible

We can often make optimal decisions without full utility information

Generalization across users

20/67Paolo Viappiani Preference Modelling and Learning

Page 21: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Adaptive Utility ElicitationUtility-based Interactive Recommender System:

Bel : belief about the user’s utility function u

Opt(Bel): optimal decision given incomplete beliefs about u

Algorithm: Adaptive Utility Elicitation

1 Repeat until Bel meets some termination condition

1 Ask user some query2 Observe user response r3 Update Bel given r

2 Recommend Opt(Bel)

Types of Beliefs

• Probabilistic Uncertainty: distribution of parameters, updated using Bayes• Strict Uncertainty: feasible region (if linear constraints: convex polytope)

21/67Paolo Viappiani Preference Modelling and Learning

Page 22: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Which Recommendations? Which Queries?

Several FrameworsThe questions:

How to represent the uncertainty over an utility function?

How to aggregate this uncertainty? Make recommendations ?

How to choose the next query? How to recommend a set, a ranking, ...

Minimax Regret Maximin BayesianKnoweldge constraints constraints prob. distribtion

representationWhich option minimax maximin expected

to recommend? regret utility utilityWhich query worst-case worst-case expected valueto ask next? regret reduction maximin improvement of information

Possible other choices: Hurwitz criterion, and others.Hybrid models: minimax regret with expected regret reduction,...

22/67Paolo Viappiani Preference Modelling and Learning

Page 23: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Minimax RegretIntuitionAdversarial game; the recommender selects the item reducing the “regret” wrt the“best” item when the uniknonw parameters are chosen by the adversary

Robust criterion for decision making under uncertainty [Savage; Kouvelis]

Showed to be effective when used for decision making under utility uncertainty [Boutilier et al., 2006] and asdriver for elicitation

Advantages

No heavy Bayesian updates

No prior assumption required

MMR computation suggests queries to ask to the user

Limitations

No account for noisy responses

Formulation of the optimization depends on the assumption about the utility

23/67Paolo Viappiani Preference Modelling and Learning

Page 24: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Minimax Regret

Assumption: a set of feasible utility functions W is given

The pairwise max regret

PMR(x, y; W ) = maxw∈W

u(y; w)− u(x; w)

The max regret

MR(x; W ) = maxy∈X

PMR(x, y; W )

The minimax regret

MMR(W ) of W and the minimax optimal item x∗W :

MMR(W ) = minx∈X

MR(x,W )

x∗W = arg minx∈X

MR(x,W )

24/67Paolo Viappiani Preference Modelling and Learning

Page 25: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example

item feature1 feature2a 10 14b 8 12c 7 16d 14 9e 15 6f 16 0

Linear utility model with normalized utility weights(w1 + w2 = 1);u(x ; w) = (1−w2)x1 +w2x2 = (x2−x1)w2 + x1

Notice: it is a 1 dimensional problem

Initially, we only know that w2 ∈ [0, 1]

PMR(a, f ; w2)= maxw2 u(f ; w2)−u(a; w2)= maxw2 6(1−w2)− 14w2 = maxw2 6− 20w2

= 6 (for w2 = 0)

PMR(a, b; w2)=maxw2 u(b; w2)−u(a; w2)<0(a dominates b; there can’t be regret in choosinga instead of b!)

PMR(a, c; w2)=maxw2 −3(1−w2)−2w2 =2(for w2 = 1)....

25/67Paolo Viappiani Preference Modelling and Learning

Page 26: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example (continued)

item feature1 feature2a 10 14b 8 12c 7 16d 14 9e 15 6f 16 0

Linear utility model with normalized utility weights(w1 + w2 = 1);u(x ; w) = (1−w2)x1 +w2x2 = (x2−x1)w2 + x1

Notice: it is a 1 dimensional problem

Computation of the pairwise regret table.

PMR(·, ·) a b c d e f MRa 0 -2 2 4 5 6 6b 2 0 4 6 7 8 8c 3 1 0 7 8 9 9d 5 3 7 0 1 2 7e 8 6 10 3 0 1 10f 14 12 16 9 6 0 16

The MMR-optimal solution is a, adversarial choiceis f , and minimax regret value is 6.

In reality no need to compute the full table (treesearch methods) [Braziunas, PhD Thesis, 2011]

Now, we want to ask a new query to improve thedecision. A very successful strategy (thoughtgenerally not optimal!) is the current solutionstrategy: ask user to compare a and f

26/67Paolo Viappiani Preference Modelling and Learning

Page 27: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

INCREMENTAL ECLITATION WITH MINIMAX REGRET

OBSERVATIONS

Si P ⊂ P ′ alors :

ΘP′ ⊂ ΘP

PMR(x, y; ΘP′) ≤ PMR(x, y; ΘP) for all x, y ∈ XMR(x; ΘP′) ≤ MR(x; ΘP) for all x ∈ XMMR(ΘP′) ≤ MMR(ΘP)

→ Adding preference statements cannot increase MMR (and oftendecreases); justification for interactive elicitation with questions.

WHICH STRATEGY TO ASK QUESTIONS?

Worst-case Minimax Regret : arg minq∈Q

maxr

MMR(X ; ΘP∪r)

Current Choice Strategy : x∗ % y∗ or x∗ - y∗?(do you prefer x∗ or y∗ ?)

27/67Paolo Viappiani Preference Modelling and Learning

Page 28: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

A graphical illustration for linear utility model (1/3)

28/67Paolo Viappiani Preference Modelling and Learning

Page 29: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

A graphical illustration for linear utility model (2/3)

29/67Paolo Viappiani Preference Modelling and Learning

Page 30: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

A graphical illustration for linear utility model (3/3)

30/67Paolo Viappiani Preference Modelling and Learning

Page 31: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Minimax Regret ComputationComputation of Pairwise Max Regret as Linear Program

Objective function: maxw∈W w · (y− x)

Usually W expressed by linear constraints such as w ·a ≥ w ·b fora % b ∈ P (a set of comparisons)

Computation of Minimax Regret

Naive approach: test all n2−n combinations of choices

Better idea: implement a search problem [Braziunas, 2011]

i=choice of recommender, j=choice of adversaryUB: upper bound on minimax regret (max regret of best solutionfound)LBi : lower bound on the max regret of option iAfter testing i against j : LBi ← max(LBi ,PMR(i, j))Whenever LBi ≥ UB: prune option i

Empirically, a small number of PMR checks is needed31/67

Paolo Viappiani Preference Modelling and Learning

Page 32: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example of Minimax Regret Computation

ExampleComplete “pairwise max regret" table

PMR(i , j) j=1 j=2 j=3 j=4 MR(i)i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 1 0 1 4i=4 3 2 3 0 3

Evaluation (0 PMR checks)PMR(i , j) j=1 j=2 j=3 j=4 LBi

i=1 0 ? ? ? 0i=2 ? 0 ? ? 0i=3 ? ? 0 ? 0i=4 ? ? ? 0 0

UB+Inf

32/67Paolo Viappiani Preference Modelling and Learning

Page 33: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example of Minimax Regret Computation

ExampleComplete “pairwise max regret" table

PMR(i , j) j=1 j=2 j=3 j=4 MR(i)i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 1 0 1 4i=4 3 2 3 0 3

Evaluation (1 PMR checks)PMR(i , j) j=1 j=2 j=3 j=4 LBi

i=1 0 1 ? ? 1i=2 ? 0 ? ? 0i=3 ? ? 0 ? 0i=4 ? ? ? 0 0

UB+Inf

33/67Paolo Viappiani Preference Modelling and Learning

Page 34: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example of Minimax Regret Computation

ExampleComplete “pairwise max regret" table

PMR(i , j) j=1 j=2 j=3 j=4 MR(i)i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 1 0 1 4i=4 3 2 3 0 3

Evaluation (3 PMR checks)PMR(i , j) j=1 j=2 j=3 j=4 LBi

i=1 0 1 2 3 3i=2 ? 0 ? ? 0i=3 ? ? 0 ? 0i=4 ? ? ? 0 0

UB3

34/67Paolo Viappiani Preference Modelling and Learning

Page 35: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example of Minimax Regret Computation

ExampleComplete “pairwise max regret" table

PMR(i , j) j=1 j=2 j=3 j=4 MR(i)i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 1 0 1 4i=4 3 2 3 0 3

Evaluation (6 PMR checks)PMR(i , j) j=1 j=2 j=3 j=4 LBi

i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 ? ? 0 ? 0i=4 ? ? ? 0 0

UB2

35/67Paolo Viappiani Preference Modelling and Learning

Page 36: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example of Minimax Regret Computation

ExampleComplete “pairwise max regret" table

PMR(i , j) j=1 j=2 j=3 j=4 MR(i)i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 1 0 1 4i=4 3 2 3 0 3

Evaluation (7 PMR checks)PMR(i , j) j=1 j=2 j=3 j=4 LBi

i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 ? 0 ? 4i=4 ? ? ? 0 0

UB2

36/67Paolo Viappiani Preference Modelling and Learning

Page 37: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Example of Minimax Regret Computation

ExampleComplete “pairwise max regret" table

PMR(i , j) j=1 j=2 j=3 j=4 MR(i)i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 1 0 1 4i=4 3 2 3 0 3

Evaluation (8 PMR checks)PMR(i , j) j=1 j=2 j=3 j=4 LBi

i=1 0 1 2 3 3i=2 2 0 2 2 2i=3 4 ? 0 ? 4i=4 3 ? ? 0 3

UB2

37/67Paolo Viappiani Preference Modelling and Learning

Page 38: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Different Aggregation Functionsx = (x1, . . . , xn), y = (y1, . . . , yn) → x % y⇔ u(x;ω) ≥ u(y;ω) Weightingparameters provide a control on:

the type of compromise sought in Multicriteria Decision Making

the attitude towards equity in Social Choice

Standard weighting parameters (weights attached to criteria)

•Weighted sum : u(x;ω) =n∑

i=1ωixi

•Weighted Tchebycheff : u(x;ω) = maxi∈[[1;n]]

{ωix∗i −xi

x∗i −x∗i}

Rank-dependent weighting parameters (weights attached to ranks)

• OWA : u(x; w) =n∑

i=1wix(i)

• Choquet : u(x; v) =n∑

i=1

[x(i) − x(i−1)

]v(X(i))

38/67Paolo Viappiani Preference Modelling and Learning

Page 39: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Application to OWA

Positive normalized weights (general case)

u(x; w) =n∑

i=1wix(i) non-linear in x but linear in w !

u(x ,w) ≥ u(y ,w) ⇐⇒ w .x↑ ≥ w .y↑

where x↑ is the vector x sorted by increasing order→preference of type x % y is equivalent to a linear inequality

Positive normalized decreasing weights (fair optimization)OWA with decreasing weights: wi > wj whenever i < j ensures thecompatibility with Pigou-Dalton transfers, i.e:∀i , j : xj > xi ,∀ε ∈ (0, xj − xi ), (x1, xi + ε, . . . , xj − ε, xn) � (x1, . . . , xn)→ add inequalities of type wi − wi+1 ≥ δ with δ > 0

39/67Paolo Viappiani Preference Modelling and Learning

Page 40: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Regret Computation for OWA

Dataset Manipulation

x↑ = (x(1), ..., x(n)) permutation of features from worst to best

PMR for OWA

maxw

w .(y↑ − x↑) (3)

s.t . 0 ≤ wi ≤ 1 ∀i (4)∑i

wi = 1 (5)

wi − wi+1 ≥ δ (6)

w .a↑ ≥ w .b↑ ∀a, b s.t. a % b ∈ P (7)

40/67Paolo Viappiani Preference Modelling and Learning

Page 41: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Application to Choquet integrals

Cv (x) = x(1)v(X(1)) +n∑

i=2

[x(i) − x(i−1)

]v(X(i))

x(i) ≤ x(i+1) for all i = 1, . . . ,n − 1 and X(i) = {j ∈ N, xj ≥ x(i)},

Example

∅ {1} {2} {3} {1,2} {1,3} {2,3} {1,2,3}v 0 0.1 0.2 0.3 0.5 0.6 0.7 1

x = (10,6,14) and y = (10,12,8)

Cv (x) = 6 + (10− 6)v({1,3}) + (14− 10)v({3}) = 9.6Cv (y) = 8 + (10− 8)v({1,2}) + (12− 10)v({2}) = 9.4

With Cv we observe that x is preferred to y .

41/67Paolo Viappiani Preference Modelling and Learning

Page 42: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Fairness and convex capacities in the Choquet integral

v is said to be convex or supermodular whenv(A ∪ B) + v(A ∩ B) ≥ v(A) + v(B) for all A,B ⊆ N,

Proposition (Chateauneuf and Tallon, 1999)When preferences are represented by a Choquet integral, thenchoosing v convex is equivalent to the following property:∀x1, x2, . . . xp ∈ Rn, ∀k ∈ {1,2, . . . ,p} and ∀i ∈ {1,2, . . . ,p}, λi ≥ 0such that

∑pi=1 λi = 1 we have:

Cv (x1) = Cv (x2) = . . . = Cv (xp)⇒ Cv (

p∑i=1

λix i ) ≥ Cv (xk )

Example (convex v and fairness)x = (18,18,0), y = (0,18,18), z = (18,0,18)t = (12,12,12) = (x + y + z)/3v convex⇒ [Cv (x) = Cv (y) = Cv (z)⇒ t % x , t % y , t % z]

42/67Paolo Viappiani Preference Modelling and Learning

Page 43: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

COMPUTATION OF PMR FOR CHOQUET: GENERAL

CASE

CAPACITY

The function v : 2N → [0, 1] is a normalized capacity if:

v(∅) = 0, v(N) = 1

v(A) ≤ v(B) for all A ⊂ B ⊆ N (monotonicity)

PMR(x, y; ΘP) = maxv

Cv(y)− Cv(x)

s.t. v(∅) = 0

v(N) = 1

v(A) ≤ v(A ∪ {i}) ∀A ⊂ N, ∀i ∈ N\ACv(a) ≥ Cv(b) ∀a, b s.t. a % b ∈ P

LP with an exponentiel number of variables and constraintsAt most 2n variables in the objective function

43/67Paolo Viappiani Preference Modelling and Learning

Page 44: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

COMPUTATION OF PMR FOR CHOQUET: GENERAL

CASE

CAPACITY

The function v : 2N → [0, 1] is a normalized capacity if:

v(∅) = 0, v(N) = 1

v(A) ≤ v(B) for all A ⊂ B ⊆ N (monotonicity)

PMR(x, y; ΘP) = maxv

Cv(y)− Cv(x)

s.t. v(∅) = 0

v(N) = 1

v(A) ≤ v(A ∪ {i}) ∀A ⊂ N, ∀i ∈ N\ACv(a) ≥ Cv(b) ∀a, b s.t. a % b ∈ P

LP with an exponentiel number of variables and constraintsAt most 2n variables in the objective function

43/67Paolo Viappiani Preference Modelling and Learning

Page 45: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Choquet integral and Möbius massesAn alternative representation in terms of the Möbius inverse:

Möbius inverse and Möbius masses

To any set-function v : 2N → R is associated m : 2N → R a mapping calledMöbius inverse, defined by:

∀A ∈ N, m(A) =∑B⊆A

(−1)|A\B|v(B), v(A) =∑B⊆A

m(B)

Coefficients m(B) for B ⊆ A are called Möbius masses.

Remark: v convex if Möbius masses positive (belief function, Shafer 1976)

Choquet integral as a function of Möbius masses

Cv (x) =∑B⊆A

m(B)∧i∈B

xi

2-additive capacity (m(B) = 0 iff |B| > 2) → capacity completely characterized by (n2 + n)/2 coefficients. Good

compromise between compacity of the model and expressivity. 44/67Paolo Viappiani Preference Modelling and Learning

Page 46: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

PMR for 2-additive Choquetm vector of variables (moebsius mass) encoding a capacity v :m = (m1, . . . ,mn,m11,m12, . . . ,m1n,m23, . . . ,m2n . . . ,mn−1n)x = (x1, . . . , xn, x11, x12, . . . , x1n, x23, . . . , x2n . . . , xn−1n), xij = xi ∧ xj

Therefore we can write Cv (x) = m.(x)

The case of a 2-additive capacity

maxm

m.(x − y) (8)

s.t .∑j∈J

mij ≥ 0, i =1,. . ., n; ∀J : {i} ⊆ J ⊆ N (9)

n∑i=1

mi +n∑i

n∑j=i+1

mij = 1 (10)

m.a ≥ m.b ∀a, b s.t. a % b ∈ P

Constraint 11 ensures that the capacity is normalized. Monotonicity (9) requires n 2n−1 constraints. Only for

2-additive capacities: efficient formulation in term of convex combination of extreme masses [Hullermeier 2012]. 45/67Paolo Viappiani Preference Modelling and Learning

Page 47: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

PMR for 2-additive Belief Functionsm vector of variables (moebsius mass) encoding a capacity v :m = (m1, . . . ,mn,m11,m12, . . . ,m1n,m23, . . . ,m2n . . . ,mn−1n)x = (x1, . . . , xn, x11, x12, . . . , x1n, x23, . . . , x2n . . . , xn−1n), xij = xi ∧ xj

Cv (x) = m.(x)

belief functions are capacities such that mi,j ≥ 0

The case of (2-additive) belief functions

maxm

m.(x − y) (11)

s.t .n∑

i=1

mi +n∑i

n∑j=i+1

mij = 1 (12)

mi ≥ 0, i =1, . . . ,n, (13)mij ≥ 0, i =1, . . . ,n, j = i + 1, . . . ,n, (14)

m.a ≥ m.b ∀a, b s.t. a % b ∈ P46/67

Paolo Viappiani Preference Modelling and Learning

Page 48: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

PREFERENCE QUERIES OF TYPE “1A0 � Λ ?”COMPARISON BETWEEN:

1A0, the fictions alternative such that 1A0i = 1 if i ∈ A and 1A0i = 0otherwise, andΛ, the fictious alternative such that Λi = λ for all i ∈ N .

OBSERVATIONS

Cv(1A0) = v(A) for all A ⊆ NCv(Λ) = λ for all λ ∈ R

CONSEQUENCES

1A0 % Λ ⇔ v(A) ≥ λ ⇔ v(B) ≥ λ for all B ⊇ A1A0 - Λ ⇔ v(A) ≤ λ ⇔ v(B) ≤ λ for all B ⊆ A

PROPOSITION 1 : COMPATIBILITY WITH PREFERENCES PUpdating [lB , uB] for all B ⊇ A (resp. B ⊆ A) at each insertion of preferences1A0 % Λ (resp. 1A0 - Λ) in P, the compatibility with P is assured.

47/67Paolo Viappiani Preference Modelling and Learning

Page 49: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

PREFERENCE QUERIES OF TYPE “1A0 � Λ ?”COMPARISON BETWEEN:

1A0, the fictions alternative such that 1A0i = 1 if i ∈ A and 1A0i = 0otherwise, andΛ, the fictious alternative such that Λi = λ for all i ∈ N .

OBSERVATIONS

Cv(1A0) = v(A) for all A ⊆ NCv(Λ) = λ for all λ ∈ R

CONSEQUENCES

1A0 % Λ ⇔ v(A) ≥ λ ⇔ v(B) ≥ λ for all B ⊇ A1A0 - Λ ⇔ v(A) ≤ λ ⇔ v(B) ≤ λ for all B ⊆ A

PROPOSITION 1 : COMPATIBILITY WITH PREFERENCES PUpdating [lB , uB] for all B ⊇ A (resp. B ⊆ A) at each insertion of preferences1A0 % Λ (resp. 1A0 - Λ) in P, the compatibility with P is assured.

47/67Paolo Viappiani Preference Modelling and Learning

Page 50: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

COMPUTATION OF PMR FOR CHOQUET

Let P be preferences statements obtained by asking only 1A0 % Λ queries.

PROPOSITION 2 : EXISTENCE OF A COMPATIBLE CAPACITY

Let v|D : 2N → [0, 1] be a partial function defined on D ⊆ 2N such that:• v|D(A) ∈ [lA, uA] for all A ∈ D• v|D(A) ≤ v|D(B) for all A,B ∈ D such that A ⊂ B

In this case, v|D can be completed in a capacity of ΘP .

PMR(x, y; ΘP) = maxv

Cv(y)− Cv(x)

s.t. v(X(i+1)) ≤ v(X(i)) ∀i ∈ {1, . . . , n− 1}v(Y(i+1)) ≤ v(Y(i)) ∀i ∈ {1, . . . , n− 1}v(X(i)) ≤ v(Y(j)) ∀i, j ∈ N s.t. X(i) ⊂ Y(j) and X(i) 6⊂ Y(j+1)

v(Y(i)) ≤ v(X(j)) ∀i, j ∈ N s.t. Y(i) ⊂ X(j) and Y(i) 6⊂ X(j+1)

uX(i)≤ v(X(i)) ≤ lX(i) ∀i∈N

uY(i)≤ v(Y(i)) ≤ lY(i)

∀i ∈ N

→ LP with at most 2(n−1) variables et 4(n−1)+2n = 6n− 4 constraints.

48/67Paolo Viappiani Preference Modelling and Learning

Page 51: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

COMPUTATION OF PMR FOR CHOQUET

Let P be preferences statements obtained by asking only 1A0 % Λ queries.

PROPOSITION 2 : EXISTENCE OF A COMPATIBLE CAPACITY

Let v|D : 2N → [0, 1] be a partial function defined on D ⊆ 2N such that:• v|D(A) ∈ [lA, uA] for all A ∈ D• v|D(A) ≤ v|D(B) for all A,B ∈ D such that A ⊂ B

In this case, v|D can be completed in a capacity of ΘP .

PMR(x, y; ΘP) = maxv

Cv(y)− Cv(x)

s.t. v(X(i+1)) ≤ v(X(i)) ∀i ∈ {1, . . . , n− 1}v(Y(i+1)) ≤ v(Y(i)) ∀i ∈ {1, . . . , n− 1}v(X(i)) ≤ v(Y(j)) ∀i, j ∈ N s.t. X(i) ⊂ Y(j) and X(i) 6⊂ Y(j+1)

v(Y(i)) ≤ v(X(j)) ∀i, j ∈ N s.t. Y(i) ⊂ X(j) and Y(i) 6⊂ X(j+1)

uX(i)≤ v(X(i)) ≤ lX(i) ∀i∈N

uY(i)≤ v(Y(i)) ≤ lY(i)

∀i ∈ N

→ LP with at most 2(n−1) variables et 4(n−1)+2n = 6n− 4 constraints.

48/67Paolo Viappiani Preference Modelling and Learning

Page 52: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

COMPUTATION OF PMR FOR CHOQUET

A closer look at the Objective Function

Remember that Cv (x) = x(1)v(X(1)) +∑n

i=2

[x(i) − x(i−1)

]v(X(i)).

Cv(y)− Cv(x) =∑

A∈{X(i) | X(i) 6=Y(i)}

−(x(i) − x(i−1))vA + (15)

∑A∈{Y(i) | Y(i) 6=X(i)}

(y(i) − y(i−1))vA + (16)

∑A∈{X(i) | X(i)=Y(i)}

(y(i) − y(i−1) − x(i) + x(i−1))vA. (17)

Objective function has to be maximized: vA will be as small aspossible for all A ∈ {X(i) | X(i) 6= Y(i)} and as large as possible for allA ∈ {Y(i) | Y(i) 6= X(i)}.

49/67Paolo Viappiani Preference Modelling and Learning

Page 53: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

COMPUTATION OF PMR FOR CHOQUET

Let P be preferences statements obtained by asking only 1A0 % Λ queries.

PMR(x, y; ΘP) = maxv

Cv(y)− Cv(x)

s.t. v(X(i+1)) ≤ v(X(i)) ∀i ∈ {1, . . . , n− 1}v(Y(i+1)) ≤ v(Y(i)) ∀i ∈ {1, . . . , n− 1}v(X(i)) ≤ v(Y(j)) ∀i, j ∈ N s.t. X(i)⊂Y(j) and X(i) 6⊂Y(j+1)

v(Y(i)) ≤ v(X(j)) ∀i, j ∈ N s.t. Y(i)⊂X(j) and Y(i) 6⊂X(j+1)

Y(i)⊂X(j), Y(i) 6⊂X(j+1), and Y(i−1) 6⊂X(j)

uX(i)≤ v(X(i)) ≤ lX(i) ∀i∈N

uY(i)≤ v(Y(i)) ≤ lY(i)

∀i ∈ N

→ LP with at most 2(n−1) variables et 3(n−1)+2n = 5n− 3 constraints.

[Benabbou et al., 2014]

50/67Paolo Viappiani Preference Modelling and Learning

Page 54: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

QUERY STRATEGIES FOR CHOQUET

SELECTION CRITERIA: WORST-CASE MINIMAX REGRET

minA⊆N

minΛ

max{MMR(X ; ΘP∪{1A0%Λ}),MMR(X ; ΘP∪{1A0-Λ})}

(myopic measure of “value of information”)

OBSERVATIONS

For all A ⊆ N :

MMR(X ; ΘP∪{1A0%Λ}) is a function of λ decreasing on [lA, uA]

MMR(X ; ΘP∪{1A0-Λ}) function of λ increasing on [lA, uA]

These two functions have the same maximumCONSEQUENCES

MMR(X ; ΘP∪{1A0%Λ}) and MMR(X ; ΘP∪{1A0-Λ}) have necessarily anintersection.Minimization of Λ can be done with dichotomic search.

51/67Paolo Viappiani Preference Modelling and Learning

Page 55: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

QUERY STRATEGIES FOR CHOQUET

OBSERVATION 1 :There are exactly 2n − 2 possible subsets of criteria A ⊆ N .

HEURISTIC

Consider only the attributes implicated in the objective PMR(x∗, y∗; ΘP) forx∗ and y∗ that are associated with minimax regret.

REMARK 2 :It is possible that no single query can reduce MMR

TIE-BREAKING

Choose A minimizing λ =lA + uA

2:

0.5 MMR(X ; ΘP∪{1A0%Λ}) + 0.5 MMR(X ; ΘP∪{1A0-Λ})

[Benabbou et al., 2014] 52/67Paolo Viappiani Preference Modelling and Learning

Page 56: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Experiment 1

10-dimensional Knapsack problem with 1000 items. Interactive elicitation with a simulated user; minimax-regret

assuming a Choquet integral, when the user is answering according to different underlying models.

53/67Paolo Viappiani Preference Modelling and Learning

Page 57: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Experiment 2

Datasets of 100 alternatives evaluated on 10 criteria and characterized by a set of performance vectors X a are

randomly generated; constructed in such a way that∑n

i=1 xai = 1 for all x ∈ X a , where a ∈ {0.5, 1, 2} so as to

obtain different types of Pareto sets.

54/67Paolo Viappiani Preference Modelling and Learning

Page 58: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Which Recommendations? Which Queries?

Several FrameworsThe questions:

How to represent the uncertainty over an utility function?

How to aggregate this uncertainty? Make recommendations ?

How to choose the next query? How to recommend a set, a ranking, ...

Minimax Regret Maximin BayesianKnoweldge constraints constraints prob. distribtion

representationWhich option minimax maximin expected

to recommend? regret utility utilityWhich query worst-case worst-case expected valueto ask next? regret reduction maximin improvement of information

Possible other choices: Hurwitz criterion, and others.Hybrid models: minimax regret with expected regret reduction,...

55/67Paolo Viappiani Preference Modelling and Learning

Page 59: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Bayesian Framework for Recommendation andElicitation

Let’s assume utility u(x ; w) parametric in w for agiven structure, for example u(x ; w)=w · xP(w) probability distribution over utility function

Expected utility of a given item xEU(x)=

∫u(x) P(w) dw

Current expected utility of best recommendation x∗

EU∗=maxx∈A EU(x); x∗=arg maxx∈A EU(x)

When a new preference is known (for instance,user prefers apples over orange), the distribution isupdated according to Bayes (Monte Carlo methods,Expectation Propagation)

(possible prior distribution)

(distribution updated after user feedback)

56/67Paolo Viappiani Preference Modelling and Learning

Page 60: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Response Models

Model the user’s cognitive ability of answering correctly to a preference query

Noiseless responses (unrealistic but often assumed in researchpapers!)

Constant error (can model distraction, e.g. clicking on the wrong icon)

Logistic error (Boltzmann distribution), a commonly used probabilisticresponse model for comparison/choice queries: “Among options in setS, which one do you prefer?”

probability of response “x is my preferred item in S”

Pr (S → x) =eγu(x)∑

y∈S eγu(y)

γ is a temperature parameter (how “noisy” is the user).

For comparison queries (“Is item1 better than item2?”)P(selecting 1st item) as a function of the difference in utility

57/67Paolo Viappiani Preference Modelling and Learning

Page 61: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

What Query to Ask Next?

The problem can be modeled as a POMDP [Boutilier, AAAI 2002],however impractical to solve for non trivial cases

Idea: ask query with highest “value”, a posteriori improvement indecision quality

In a Bayesian approach (Myopic) Expected Value of Information

EVOIθ(q) =∑r∈R

Pθ(r)EU∗θ|r − EU∗θ

where R is the set of possible responses (answers); θ is the currentbelief distribution and θ|r the posterior and Pθ(r) the prior probability ofa given response.

Ask query q∗ = arg max EVOIθ(q) with highest EVOI

In non-Bayesian setting, one can use non probabilistic measures ofdecision improvement (for example, worst-case regret reduction, ...)

58/67Paolo Viappiani Preference Modelling and Learning

Page 62: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Efficient Computation of QueriesOptimal Query Sets and Recommendation Sets

How to optimize VoI depends on the query type, in general too many possiblequeries to iterate over them

For comparison queries there are n(n−1)2 possible queries

For choice queries of size k there are(n

k

)candidate queries

For bound queries, continuous space

Optimal Query Sets and Recommendation Sets

Characterization of the myopically optimal choice query[Viappiani and Boutilier, 2010]

Tight connection with problem of generating a set of recommendations

The optimization problem is submodular→ greedy optimization givesstrong guarantees

Lazy evaluation techniques are computationally very efficient (<1second for large datasets; naive methods several hours...)

59/67Paolo Viappiani Preference Modelling and Learning

Page 63: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Experimental Results (Bayesian Elicitation)

We know that the myopic value of information of the queries posed by thegreedy strategies are close to the optimum. Empirical results show that theyare also effective when evaluated in an iterative process; the recommendeditem converges to the “true” best item in few cycles.

60/67Paolo Viappiani Preference Modelling and Learning

Page 64: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Application: Active Collaborative Filtering

Incremental Elicitation of Rating Profiles

Recommender systems such as Netflix asks you to rate moremovies, in order to improve their preference models andprovide better recommendations in the future.This can be seen as an implicit “query” asking for yourpreferred movies among the ones presented.

61/67Paolo Viappiani Preference Modelling and Learning

Page 65: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

Other Approaches

Point-wise methods (Support Vector Machines)Sorting / Ordered classification problemsLearning preference relations (qualitative preferences)Learning CP networksRelational learningRankings: neighbour models, probabilistic approaches(Mallows, Babington-Smith),...

62/67Paolo Viappiani Preference Modelling and Learning

Page 66: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

The Road Ahead

Preference learning must deal with “biased” decisionmakers: more work at the intersection of behavioraldecision theory and principled mathematical methodsLearning/elicitation wrt Prospect TheorySocial context of personalizationUtility-based evaluation of “contagion” effect in socialnetworksSequential decision problems (MDPs / reinforcementlearning) and sequential optimization of elicitation(optimization of VOI for long horizon)

63/67Paolo Viappiani Preference Modelling and Learning

Page 67: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

References 1 (Fundamentals: Preference Modeling)

L. J. Savage, The Foundations of Statistics, Wiley, New York, 1954.

P. Kouvelis and G. Yu, Robust Discrete Optimization and Its Applications, Kluwer,Dordrecht, 1997.

Roberts F.S, Measurement theory, with applications to Decision Making, Utilityand the Social Sciences, Addison-Wesley, Boston, 1979.

Roubens M., Vincke Ph., Preference Modeling, Springer Verlag, Berlin, 1985.

Fishburn P.C., Interval Orders and Interval Graphs, J. Wiley, New York, 1985.

Fodor J., Roubens M., Fuzzy preference modelling and multicriteria decisionsupport, Kluwer Academic, Dordrecht, 1994.

Pirlot M., Vincke Ph., Semi Orders, Kluwer Academic, Dordrecht, 1997.

Fishburn P.C., Preference structures and their numerical representations,Theoretical Computer Science, vol. 217, 359-383, 1999.

Öztürk M., Tsoukiàs A., Vincke Ph., Preference Modelling, in M. Ehrgott, S.Greco, J. Figueira (eds.), State of the Art in Multiple Criteria Decision Analysis,Springer Verlag, Berlin, 27 - 72, 2005.

64/67Paolo Viappiani Preference Modelling and Learning

Page 68: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

References 2

Prefererence Representation and Languages

Craig Boutilier, Ronen I. Brafman, Carmel Domshlak, Holger H. Hoos, DavidPoole. CP-nets: A Tool for Representing and Reasoning with Conditional CeterisParibus Preference Statements. JAIR 21: 135-191 (2004)

Souhila Kaci. Working with Preferences: Less Is More. Cognitive Technologies,Springer 2011.

Behavioral Decision Theory and Biases

Kahneman, Daniel; Paul Slovic and Amos Tversky. Judgment under Uncertainty:Heuristics and Biases, Cambridge University Press, 1982.

Gilovich, Thomas and Dale Griffin, Daniel Kahneman. Heuristics and biases:The psychology of intuitive judgment, Cambridge University Press, 2002.

John W. Payne, James R. Bettman and Eric J. Johnson. The Adaptive DecisionMaker, Cambridge University Press, 1993.

Daniel Kahneman. Thinking, Fast and Slow. Farrar, Straus and Giroux, 2011.

65/67Paolo Viappiani Preference Modelling and Learning

Page 69: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

References 3 (Incremental Utility Elicitation/Learning)Urszula Chajewska, Daphne Koller, Ronald Parr. Making Rational DecisionsUsing Adaptive Utility Elicitation. AAAI 2000: 363-369.Craig Boutilier. A POMDP Formulation of Preference Elicitation Problems.AAAI/IAAI 2002: 239-246.C. Boutilier, R. Patrascu, P. Poupart, D. Schuurmans. Constraint-basedoptimization and utility elicitation using the minimax decision criterion. ArtificialIntelligence 170(8- 9): 686-713 (2006).Paolo Viappiani, Craig Boutilier. Regret-based optimal recommendation sets inconversational recommender systems. RecSys 2009: 101-108.Paolo Viappiani, Craig Boutilier. Optimal Bayesian Recommendation Sets andMyopically Optimal Choice Query Sets. NIPS 2010: 2352-2360.Darius Braziunas, Craig Boutilier. Assessing regret-based preference elicitationwith the UTPREF recommendation system. ACM EC 2010: 219-228.D. Braziunas. Decision-theoretic elicitation of generalized additive utilities, Ph.D.dissertation, University of Toronto, 2011.A. F. Tehrani, W. Cheng, K. Dembczynski, E. Hullermeier. Learning monotonenonlinear models using the Choquet integral. Machine Learning 89(1-2):183-211 (2012)Paolo Viappiani, Christian Kroer. Robust Optimization of Recommendation Setswith the Maximin Utility Criterion. ADT 2013: 411-424.Nawal Benabbou, Patrice Perny, Paolo Viappiani. Incremental Elicitation ofChoquet Capacities for Multicriteria Decision Making. ECAI 2014: 87-92. 66/67

Paolo Viappiani Preference Modelling and Learning

Page 70: Preference Modelling and Learningdmrsworkshop.inf.unibz.it/2014/slides/Paolo-Viappiani-DMRS2014.pdf · Preference Elicitation / Learning Are We Rational Decision Makers? NO. Human

Languages for PreferencesUtility-based Representations

Preference Elicitation / Learning

Standard vs Automated ElicitationMinimax-RegretBayesian ApproachesDiscussion and Future Works

References 4 (Applications, Intelligent User Interfaces)

Bart Peintner, Paolo Viappiani, Neil Yorke-Smith. Preferences in InteractiveSystems: Technical Challenges and Case Studies. AI Magazine 29(4): 13-24(2008).

Pearl Pu, Boi Faltings, Li Chen, Jiyong Zhang, Paolo Viappiani. UsabilityGuidelines for Product Recommenders Based on Example Critiquing Research.Recommender Systems Handbook 2011: 511-545.

Paolo Viappiani, Boi Faltings, Pearl Pu. Preference-based Search usingExample-Critiquing with Suggestions. JAIR 27: 465-503 (2006).

Krzysztof Z. Gajos, Daniel S. Weld, Jacob O. Wobbrock. Automaticallygenerating personalized user interfaces with Supple. Artificial Intelligence174(12-13): 910-950 (2010).

Markus Stolze, Michael Strobel. Recommending as Personalized Teaching -Towards Credible Needs-based eCommerce Recommender Systems inDesigning personalized user experiences in eCommerce, pages 293-313 (2004).

Myers, K.; Berry, P.; Blythe, J.; Conley, K.; Gervasio, M.; McGuinness, D.; Morley,D.; Pfeffer, A.; Pollack, M.; and Tambe, M. 2007. An Intelligent Personal Assistantfor Task and Time Management. AI Magazine 28(2): 47-61.

67/67Paolo Viappiani Preference Modelling and Learning