Controlling Fairness and Bias in Dynamic Learning-to-Rank
ACM SIGIR 2020 Marco Morik*†, Ashudeep Singh*‡, Jessica Hong‡, Thorsten Joachims‡
† TU Berlin, ‡ Cornell University
Controlling Fairness and Bias in Dynamic Learning-to-Rank ACM SIGIR - - PowerPoint PPT Presentation
Controlling Fairness and Bias in Dynamic Learning-to-Rank ACM SIGIR 2020 Marco Morik * , Ashudeep Singh * , Jessica Hong , Thorsten Joachims TU Berlin, Cornell University Dynamic Learning-to-Rank +1 +1 +1 +1 +1 1 2
ACM SIGIR 2020 Marco Morik*†, Ashudeep Singh*‡, Jessica Hong‡, Thorsten Joachims‡
† TU Berlin, ‡ Cornell University
Update Update Update
User 1 User 2 User 3 +1 +1 +1 +1 +1 +1 +1 +1
1 2 3 4 5 6
Candidate Set for query x
Position Bias
49% 51%
User Distribution
Left Leaning Right Leaning
1 2 3 4 5 6
Gleft Gright Ranking by true average relevance leads to unfair rankings.
Prefer Gright news articles. Prefer Gleft news articles.
Probability Ranking Principle [Robertson, 1977]: Rank documents by probability of relevance → 𝑧∗. Maximizes utility for virtually any measure U of ranking quality 𝑧∗ ≔ argmax𝑧 U 𝑧|𝑦
Definition: Exposure 𝑓
𝑘 is the probability a users
Exposure of Group: 𝐹𝑦𝑞 𝐻 𝑦 =
𝑘∈G
𝑓
𝑘
How to estimate?
Fang et al. 2019]
Disparate exposure allocation: A small difference in average relevance, leads to a large difference in average exposure!
Relevance = 51% Relevance = 49%
Gleft Gright
0.39 0.71 0.49 0.51
0.02 difference in expected relevance. 0.32 difference in expected exposure.
0.5 1 Exposure(j) = 1/log(1+j)
[Singh & Joachims. Fairness of Exposure in Rankings. KDD 2018]
Goal Exposure is allocated based on relevance
𝐹𝑦𝑞 𝐻 𝑦 = 𝑔 𝑆𝑓𝑚 𝐻|𝑦 The expected impact (e.g. clickthrough rate) is allocated based on merit. 𝐽𝑛𝑞 𝐻 𝑦 = 𝑔(𝑆𝑓𝑚 𝐻 𝑦 ) For the position bias model, 𝐽𝑛𝑞 𝑒 𝑦 = 𝐹𝑦𝑞 𝑒|𝑦 𝑆𝑓𝑚 𝑒 𝑦 Constraint Make exposure proportional to relevance (per group) 𝐹𝑦𝑞(𝐻0|𝑦) 𝐹𝑦𝑞 𝐻1|𝑦 = 𝑆𝑓𝑚(𝐻0|𝑦) 𝑆𝑓𝑚 𝐻1 𝑦 . Make the expected impact proportional to relevance (per group) 𝐽𝑛𝑞(𝐻0|𝑦) 𝐽𝑛𝑞 𝐻1|𝑦 = 𝑆𝑓𝑚(𝐻0|𝑦) 𝑆𝑓𝑚(𝐻1|𝑦) . Disparity Measure 𝐸𝐹 𝐻0, 𝐻1 = 𝐹𝑦𝑞(𝐻0|𝑦) 𝑆𝑓𝑚(𝐻0|𝑦) − 𝐹𝑦𝑞 𝐻1|𝑦 𝑆𝑓𝑚(𝐻1|𝑦) . 𝐸𝐽 𝐻0, 𝐻1 = 𝐽𝑛𝑞 𝐻0|𝑦 𝑆𝑓𝑚 𝐻0|𝑦 − 𝐽𝑛𝑞 𝐻1|𝑦 𝑆𝑓𝑚 𝐻1|𝑦 .
Does not satisfy Fairness of Exposure or Fairness of Impact. Relevance = 51% Relevance = 49%
0.39 0.71 0.49 0.51
0.5 1
Exposure(j) = 1/log(1+j)
Sequentially present rankings to users that ❑ Maximize Expected User Utility 𝔽 𝑉 𝑦 ❑ Ensure Unfairness 𝐸𝜐 goes to 0 with 𝜐.
FairCo: Ranking at time 𝜐 𝜏𝜐 = argsort𝑒∈ 𝑆 𝑒 𝑦 + 𝜇 errτ 𝑒
rate of 𝒫
1 𝜐 .
𝑆(𝑒).
𝑆 𝑒 𝑦 for personalization.
Proportional Controller: Linear feedback control system where correction is proportional to the error. 𝑆 𝑒 𝑦 : Estimated Conditional Relevance 𝑓𝑠𝑠
𝜐 𝑒 = 𝜐 − 1 max 𝐻𝑗 (
𝐸𝜐
𝐹(𝐻𝑗, 𝐻(𝑒)))
𝜇 > 0
𝑑𝑢 𝑒 : Click on 𝑒 at time 𝑢. 𝑞𝑢 𝑒 : Position bias at the position of 𝑒. [Joachims et al., 2017]
Experimental Evaluation
A user’s relevance is a function of their polarity and the news article’s polarity, and their openness.
Prefer Gright news articles. Prefer Gleft news articles.
Gleft 𝜍𝑒 < 0 Gright 𝜍𝑒 ≥ 0
Each news source in the dataset has a polarity assigned 𝜍𝑒 ∈ [−1, 1]. Sample user 𝑣𝑢 is drawn with a polarity parameter 𝜍𝑣𝑢 ∈ [−1, 1] and an openness parameter 𝑝𝑢 ∈ (0.05, 0.55).
Goal: Present rankings to a sequence of users to maximize their utility while providing fair exposure to the news articles relative to their average relevance over the user population.
FairCo keeps the Unfairness low for any amount of head start. Click count based ranking converges to unfair rankings due to the initial bias.
FairCo converges to fair ranking for all user distributions. Trades off utility for fairness when there is an imbalance in user distribution.
Selection Bias Fairness
𝑆𝑥 𝑒 𝒚𝑢 − Relevance of document 𝑒 for query 𝒚𝒖.
squared loss (with no position bias).
𝑆𝑥: Output of a Neural Network with weights 𝑥. 𝑑𝑢 𝑒 : Click on 𝑒 at time 𝑢. 𝑞𝑢 𝑒 : Position bias at position of 𝑒.
Groups: MGM, Warner Bros, Paramount, 20th Century Fox, Columbia.
Goal: Present ranking to each user 𝑣𝜐 to maximize NDCG while making sure the production companies receive fair share of exposure relative to the average relevance of their movies.
Exposure Unfairness Impact Unfairness Personalized Rankings achieve high utility (NDCG), while reducing Unfairness to 0 with 𝜐.
Marco Morik*†, Ashudeep Singh*‡, Jessica Hong‡, Thorsten Joachims‡
† TU Berlin, ‡ Cornell University
ACM SIGIR 2020