fairness-aware classifier with prejudice remover regularizer

36
Fairness-aware Classifier with Prejudice Remover Regularizer Toshihiro Kamishima * , Shotaro Akaho * , Hideki Asoh * , and Jun Sakuma *National Institute of Advanced Industrial Science and Technology (AIST), Japan University of Tsukuba, Japan; and Japan Science and Technology Agency ECMLPKDD 2012 (22nd ECML and 15th PKDD Conference) @ Bristol, United Kingdom, Sep. 24-28, 2012 1 START

Upload: toshihiro-kamishima

Post on 29-Nov-2014

698 views

Category:

Technology


2 download

DESCRIPTION

Fairness-aware Classifier with Prejudice Remover Regularizer Proceedings of the European Conference on Machine Learning and Principles of Knowledge Discovery in Databases (ECMLPKDD), Part II, pp.35-50 (2012) Article @ Official Site: http://dx.doi.org/10.1007/978-3-642-33486-3_3 Article @ Personal Site: http://www.kamishima.net/archive/2012-p-ecmlpkdd-print.pdf Handnote: http://www.kamishima.net/archive/2012-p-ecmlpkdd-HN.pdf Program codes : http://www.kamishima.net/fadm/ Conference Homepage: http://www.ecmlpkdd2012.net/ Abstract: With the spread of data mining technologies and the accumulation of social data, such technologies and data are being used for determinations that seriously affect individuals' lives. For example, credit scoring is frequently determined based on the records of past credit data together with statistical prediction techniques. Needless to say, such determinations must be nondiscriminatory and fair in sensitive features, such as race, gender, religion, and so on. Several researchers have recently begun to attempt the development of analysis techniques that are aware of social fairness or discrimination. They have shown that simply avoiding the use of sensitive features is insufficient for eliminating biases in determinations, due to the indirect influence of sensitive information. In this paper, we first discuss three causes of unfairness in machine learning. We then propose a regularization approach that is applicable to any prediction algorithm with probabilistic discriminative models. We further apply this approach to logistic regression and empirically show its effectiveness and efficiency.

TRANSCRIPT

Page 1: Fairness-aware Classifier with Prejudice Remover Regularizer

Fairness-aware Classifierwith Prejudice Remover Regularizer

Toshihiro Kamishima*, Shotaro Akaho*, Hideki Asoh*, and Jun Sakuma†

*National Institute of Advanced Industrial Science and Technology (AIST), Japan†University of Tsukuba, Japan; and Japan Science and Technology Agency

ECMLPKDD 2012 (22nd ECML and 15th PKDD Conference)@ Bristol, United Kingdom, Sep. 24-28, 2012

1START

Page 2: Fairness-aware Classifier with Prejudice Remover Regularizer

Overview

2

Fairness-aware Data Miningdata analysis taking into account potential issues offairness, discrimination, neutrality, or independence

fairness-aware classification, regression, or clusteringdetection of unfair events from databasesfairness-aware data publication

Examples of Fairness-aware Data Mining Applicationsexclude the influence of sensitive information, such as gender or race, from decisions or predictionsrecommendations enhancing their neutralitydata analysis while excluding the influence of uninteresting information

Page 3: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

3

Applicationsfairness-aware data mining applications

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, fairness-aware data publication

Conclusion

Page 4: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

4

Applicationsapplications of fairness-aware data mining

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, situation testing, fairness-aware data publication

Conclusion

Page 5: Fairness-aware Classifier with Prejudice Remover Regularizer

Elimination of Discrimination

5

Due to the spread of data mining technologies...Data mining is being increasingly applied for serious decisionsex. credit scoring, insurance rating, employment applicationAccumulation of massive data enables to reveal personal informationex. demographics can be inferred from the behavior on the Web

excluding the influence of these socially sensitive informationfrom serious decisions

information considered from the viewpoint of social fairnessex. gender, religion, race, ethnicity, handicap, political convictioninformation restricted by law or contractsex. insider information, customers’ private information

Page 6: Fairness-aware Classifier with Prejudice Remover Regularizer

Filter Bubble

6

[TED Talk by Eli Pariser]

The Filter Bubble ProblemPariser posed a concern that personalization technologies narrow and bias the topics of information provided to people

http://www.thefilterbubble.com/

Friend Recommendation List in Facebook

To fit for Pariser’s preference, conservative people are eliminated from his recommendation list, while this fact is not notified to him

Page 7: Fairness-aware Classifier with Prejudice Remover Regularizer

Information Neutral Recommender System

7

Information Neutral Recommender System

ex. A system enhances the neutrality in terms of whether conservative or progressive, but it is allowed to make biased recommendations in terms of other viewpoints, for example, the birthplace or age of friends

enhancing the neutrality from a viewpoint specified by a userand other viewpoints are not considered

[Kamishima+ 12]

The Filter Bubble Problem

fairness-aware data mining techniques

Page 8: Fairness-aware Classifier with Prejudice Remover Regularizer

Non-Redundant Clustering

8

[Gondek+ 04]

Simple clustering methods find two clusters: one contains only faces, and the other contains faces with shoulders

Data analysts consider this clustering is useless and uninteresting

A non-redundant clustering method derives more useful male and female clusters, which are independent of the above clusters

non-redundant clustering : find clusters that are as independent from a given uninteresting partition as possible

a conditional information bottleneck method,which is a variant of an information bottleneck method

clustering facial images

Page 9: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

9

Applicationsapplications of fairness-aware data mining

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, situation testing, fairness-aware data publication

Conclusion

Page 10: Fairness-aware Classifier with Prejudice Remover Regularizer

Difficulty in Fairness-aware Data Mining

10

US Census Data : predict whether their income is high or low

Male FemaleHigh-Income 3,256 590Low-income 7,604 4,831

fewer

Occam’s Razor : Mining techniques prefer simple hypothesis

Minor patterns are frequently ignoredand thus minorities tend to be treated unfairly

Females are minority in the high-income class# of High-Male data is 5.5 times # of High-Female dataWhile 30% of Male data are High income, only 11% of Females are

[Calders+ 10]

Page 11: Fairness-aware Classifier with Prejudice Remover Regularizer

Red-Lining Effect

11

US Census Data samples The baseline CV score is 0.19

Incomes are predicted by a naïve-Bayes classifier trained from data containing all sensitive and non-sensitive features

The CV score increases to 0.34, indicating unfair treatmentsEven if a feature, gender, is excluded in the training of a classifier

improved to 0.28, but still being unfairer than its baseline

Calders-Verwer discrimination score (CV score)Pr[ High-income | Male ] - Pr[ High-income | Female ]

the difference between conditional probabilities of advantageous decisions for non-protected and protected members. The larger score indicates the unfairer decision.

red-lining effect : Ignoring sensitive features is ineffective against the exclusion of their indirect influence

[Calders+ 10]

Page 12: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

12

Applicationsapplications of fairness-aware data mining

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, situation testing, fairness-aware data publication

Conclusion

Page 13: Fairness-aware Classifier with Prejudice Remover Regularizer

Variables

13

a numerical feature vectorfeatures other than a sensitive featurenon-sensitive, but may correlate with S

non-sensitive featuresX

objective variableY

sensitive feature

S

a binary feature variable {0, 1}socially sensitive informationex., gender or religion

a binary class variable {0, 1}a result of serious decisionex., whether or not to allow credit

Page 14: Fairness-aware Classifier with Prejudice Remover Regularizer

Prejudice

14

Direct Prejudice

Indirect Prejudice

Latent Prejudice

Prejudice : the statistical dependences of an objective variable or non-sensitive features on a sensitive feature

statistical dependence of an objective variable on a sensitive featurebringing red-lining effect

a clearly unfair state that a prediction model directly depends on a sensitive featureimplying the conditional dependence between Y and S given X

statistical dependence of non-sensitive features on a sensitive featurecompletely excluding sensitive information

Y ⊥⊥ S | X/

Y ⊥⊥ S | φ/

X ⊥⊥ S | φ/

Page 15: Fairness-aware Classifier with Prejudice Remover Regularizer

D

�Pr[X, S]�M[Y |X, S]

�Pr[Y,X, S]=

Fairness-aware Classification

15

Pr[X, S]M̂[Y |X, S]

P̂r[Y,X, S]

=

True Distribution Estimated Distribution

sample

learn

�Pr[X, S]�M†[Y |X, S]

�Pr†[Y,X, S]

=

True Fair Distribution

Pr[X, S]M̂†[Y |X, S]

P̂r†[Y,X, S]

=

Estimated Fair Dist.

learn

no indirectprejudice

no indirectprejudicesimilar

approximate

approximate

TrainingData

Page 16: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

16

Applicationsapplications of fairness-aware data mining

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, situation testing, fairness-aware data publication

Conclusion

Page 17: Fairness-aware Classifier with Prejudice Remover Regularizer

Logistic Regressionwith Prejudice Remover Regularizer

17

modifications of a logistic regression modeladd ability to adjust distribution of Y depending on the value of Sadd a constraint of a no-indirect-prejudice condition

add ability to adjust distribution of Y depending on the value of S

Multiple logistic regression models are built separately, and each of these models corresponds to each value of a sensitive featureWhen predicting scores, a model is selected according to the value of a sensitive feature

weights depending on the value of S

Pr[Y =1|x, S=0] = sigmoid(x�w0)Pr[Y =1|x, S=1] = sigmoid(x�w1)

Page 18: Fairness-aware Classifier with Prejudice Remover Regularizer

� ln Pr({(y,x, s)}; �) + � R({(y,x, s)}, �) +�

2���2

2

No Indirect Prejudice Conditon

18

L2 regularizeravoiding

over fitting

regularizationparameter λ

samples ofsensitivefeatures

Prejudice Remover Regularizerthe smaller value

more strongly constraintsthe independence between Y and S

samples ofnon-sensitive

features

samples ofobjectivevariables model

parameter

η : fairnessparameter

the larger valuemore enforces

the fairness

add a constraint of a no-indirect-prejudice condition

Page 19: Fairness-aware Classifier with Prejudice Remover Regularizer

Y �{0,1}

(x,s)

M̂[y|x, s;�] lnP̂r[y|s;�]

�s P̂r[y|s;�]

Prejudice Remover Regularizer

19

Prejudice Remover Regularizermutual information between Y (objective variable) and S (sensitive feature)

expectation over X and Sis replaced with

the summation over samples

no-indirect-prejudice condition = independence between Y and S

Y �{0,1}

X,S

Pr[X, S] M̂[Y |X, S] lnP̂r[Y, S]

P̂r[S]P̂r[Y ]

Page 20: Fairness-aware Classifier with Prejudice Remover Regularizer

Y �{0,1}

(x,s)

M̂[y|x, s;�] lnP̂r[y|s;�]

�s P̂r[y|s;�]

Computing Mutual Information

20

This distribution can be derived by marginalizing over X

approximate by the sample mean over X for each pair of y and s

�Dom(x) M̂[y|x, s; �]P̂r[x]dx

But this is computationally heavy...

�x�D M̂[y|x, s; �]

Limitation: This technique is applicable only if both Y and S are discrete

Page 21: Fairness-aware Classifier with Prejudice Remover Regularizer

Calders-Verwer’s 2 Naïve Bayes

21

Y

S X

Naïve Bayes Calders-Verwer Two Naïve Bayes (CV2NB)

S and X are conditionally independent given Y

non-sensitive features X are mutually conditionally independent given Y and S

Y

S X

[Calders+ 10]

✤ It is as if two naïve Bayes classifiers are learned depending on each value of the sensitive feature; that is why this method was named by the 2-naïve-Bayes

Unfair decisions are modeled by introducingthe dependence of X on S as well as on Y

Page 22: Fairness-aware Classifier with Prejudice Remover Regularizer

while CVscore > 0 if # of data classified as “1” < # of “1” samples in original data then increase M[Y=+, S=-], decrease M[Y=-, S=-] else increase M[Y=-, S=+], decrease M[Y=+, S=+] reclassify samples using updated model M[Y, S]

M̂[Y, S] M̂†[Y, S]

Calders-Verwer’s Two Naïve Bayes

22

keep the updated distribution similar to the empirical distribution

update the joint distribution so that its CV score decreases

[Calders+ 10]

P̂r[Y, X, S] = M̂[Y, S]�

i P̂r[Xi|Y, S]

estimated model fair estimated modelfair

parameters are initialized by the corresponding empirical distributions

M[Y, S] is modified so as to improve the fairness

Page 23: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

23

Applicationsapplications of fairness-aware data mining

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, situation testing, fairness-aware data publication

Conclusion

Page 24: Fairness-aware Classifier with Prejudice Remover Regularizer

Experimental Conditions

24

Calders & Verwer’s Test DataAdult / Census Income @ UCI RepositoryY : a class representing whether subject’s income is High or LowS : a sensitive feature representing whether subject’s genderX : non-sensitive features, all features are discretized, and 1-of-K representation is used for logistic regression# of samples : 16281

MethodsLRns : logistic regression without sensitive featuresNBns : naïve Bayes without sensitive featuresPR : our logistic regression with prejudice remover regularizerCV2NB : Calders-Verwer’s two-naïve-Bayes

Other ConditionsL2 regularization parameter λ = 1five-fold cross validation

Page 25: Fairness-aware Classifier with Prejudice Remover Regularizer

Evaluation Measure

AccuracyHow correct are predicted classes?the ratio of correctly classified sample

NMI (normalized mutual information)How fair are predicted classes?mutual information between a predicted class and a sensitive feature, and it is normalized into the range [0, 1]

25

NMI =I(Y ; S)�H(Y )H(S)

Page 26: Fairness-aware Classifier with Prejudice Remover Regularizer

0.08

0.06

0.04

0.02

0.010 10 20 30

NBnsLRnsCV2NBPR

0.76

0.78

0.80

0.82

0.84

0.86

0 10 20 30

Experimental Results

26

Our method PR (Prejudice Remover) could make fairer decisions than pure LRns (logistic regression) and NBns (naïve Bayes)PR could make more accurate prediction than NBns or CV2NB

Accuracy Fairness (NMI between S & Y)

high

-acc

urac

y

faire

r

fairness parameter η : the lager value more enhances the fairness

accuracy NMIfairness parameter η

CV2NB achieved near-zero NMI, but PR could NOT achieve it

Page 27: Fairness-aware Classifier with Prejudice Remover Regularizer

Synthetic Data

27

xi = ✏i

✏i ⇠ N (0, 1)

si 2 {0, 1} ⇠ DiscreteUniform

sensitivefeature

non-sensitivefeature

wi =

(1 + ✏i, if si = 1

�1 + ✏i, otherwise

independentfrom si

depend on si

objectivevariable yi =

(1, if xi + wi < 0

0, otherwise

both xi and wiequally influence

an objective variable

Why did our prejudice remover fail tomake a fairer prediction than that made by CV2NB?

Page 28: Fairness-aware Classifier with Prejudice Remover Regularizer

Analysis on Synthetic Data Results

28

M[ Y | X, S=0 ]M[ Y | X, S=0 ]M[ Y | X, S=0 ] M[ Y | X, S=0 ]M[ Y | X, S=0 ]M[ Y | X, S=0 ]

X W bias X W biasη=0 11.3 11.3 -0.0257 11.3 11.4 0.0595η=150 55.3 -53.0 -53.6 56.1 54.1 53.6

When η=0, PR regularizer doesn’t affect weights, and these weights for X and W are almost equalWhen η=150, absolutes of weights for X are larger than those for W

Prejudice Remover can consider the differences among individual features, but CV2NB can improve fairness more drastically

Prejudice Remover

CV2NB

PR ignores features depending on S if the influences to Y are equal

CV2NB treats all features equally

Page 29: Fairness-aware Classifier with Prejudice Remover Regularizer

Outline

29

Applicationsapplications of fairness-aware data mining

Difficulty in Fairness-aware Data MiningCalders-Verwer’s discrimination score, red-lining effect

Fairness-aware Classificationfairness-aware classification, three types of prejudices

Methodsprejudice remover regularizer, Calders-Verwer’s 2-naïve-Bayes

Experimentsexperimental results on Calders&Verwer’s data and synthetic data

Related Workprivacy-preserving data mining, detection of unfair decisions, explainability, situation testing, fairness-aware data publication

Conclusion

Page 30: Fairness-aware Classifier with Prejudice Remover Regularizer

Relation toPrivacy-Preserving Data Mining

30

indirect prejudicethe dependency between a objective Y and a sensitive feature S

from the information theoretic perspective,mutual information between Y and S is non-zero

from the viewpoint of privacy-preservation,leakage of sensitive information when an objective variable is known

differences from PPDMintroducing randomness is occasionally inappropriate for severe decisions, such as job applicationdisclosure of identity isn’t problematic generally

Page 31: Fairness-aware Classifier with Prejudice Remover Regularizer

Finding Unfair Association Rules

31

[Pedreschi+ 08, Ruggieri+ 10]

a-protection : considered as unfair if there exists association rules whose elift is larger than aex: (b) isn’t a-protected if a = 2, because elift = conf(b) / conf(a) = 3

ex: association rules extracted from German Credit Data Set

extended lift (elift)

(a) city=NYC ⇒ class=bad (conf=0.25) 0.25 of NY residents are denied their credit application(b) city=NYC & race=African ⇒ class=bad (conf=0.75) 0.75 of NY residents whose race is African are denied their credit application

elift =

the ratio of the confidence of a rule with additional conditionto the confidence of a base rule

They proposed an algorithm to enumerate rules that are not a-protected

conf(A ^ B ) C)conf(A ) C)

Page 32: Fairness-aware Classifier with Prejudice Remover Regularizer

Explainability

32

conditional discrimination : there are non-discriminatory cases, even if distributions of an objective variable depends on a sensitive feature

[Zliobaite+ 11]

sensitive featuregender

male / female

objective variableacceptance ratio

accept / not accept

explainable featureprogram

medicine / computer

Because females tend to apply to a more competitive program, females are more frequently rejectedSuch difference is explainable and is considered as non-discriminatory

more females apply to medicinemore males apply to computer

ex : admission to a university

medicine : low acceptancecomputer : high acceptance

Page 33: Fairness-aware Classifier with Prejudice Remover Regularizer

Situation Testing

33

Situation Testing : When all the conditions are same other than a sensitive condition, people in a protected group are considered as unfairly treated if they received unfavorable decision

people ina protected group

They proposed a method for finding unfair treatments by checking the statistics of decisions in k-nearest neighbors of data points in a protected groupCondition of situation testing isPr[ Y | X, S=a ] = Pr[ Y | X, S=b ] ∀ XThis implies the independence between S and Y

[Luong+ 11]

k-nearest neighbors

Page 34: Fairness-aware Classifier with Prejudice Remover Regularizer

Fairness-aware Data Publishing

34

data owner

vendor (data user)

loss functionrepresenting utilities

for the vendor

original data

archtypedata representation so as to maximize vendor’s utilityunder the constraints to guarantee fairness in analysisfair decisions

[Dwork+ 11]

They show the conditions that these archtypes should satisfyThis condition implies that the probability of receiving favorable decision is irrelevant to belonging to a protected group

Page 35: Fairness-aware Classifier with Prejudice Remover Regularizer

Conclusion

35

Contributionsthe unfairness in data mining is formalized based on independencea prejudice remover regularizer, which enforces a classifier's independence from sensitive informationexperimental results of logistic regressions with our prejudice remover

Future Workimprove the computation of our prejudice remover regularizerfairness-aware classification method for generative modelsanother type of fairness-aware mining task

Socially Responsible MiningMethods of data exploitation that do not damage people’s lives, such as fairness-aware data mining, PPDM, or adversarial learning, together comprise the notion of socially responsible mining, which it should become an important concept in the near future.

Page 36: Fairness-aware Classifier with Prejudice Remover Regularizer

Program Codes and Data Sets

Fairness-Aware Data Mininghttp://www.kamishima.net/fadm

Information Neutral Recommender Systemhttp://www.kamishima.net/inrs

AcknowledgementsWe wish to thank Dr. Sicco Verwer for providing detail information about their work and program codesWe thank the PADM2011 workshop organizers and anonymous reviewers for their valuable commentsThis work is supported by MEXT/JSPS KAKENHI Grant Number 16700157, 21500154, 22500142, 23240043, and 24500194, and JST PRESTO 09152492

36