ShaRP: Explaining Rankings with Shapley Values
CoRR(2024)
摘要
Algorithmic decisions in critical domains such as hiring, college admissions,
and lending are often based on rankings. Because of the impact these decisions
have on individuals, organizations, and population groups, there is a need to
understand them: to know whether the decisions are abiding by the law, to help
individuals improve their rankings, and to design better ranking procedures.
In this paper, we present ShaRP (Shapley for Rankings and Preferences), a
framework that explains the contributions of features to different aspects of a
ranked outcome, and is based on Shapley values. Using ShaRP, we show that even
when the scoring function used by an algorithmic ranker is known and linear,
the weight of each feature does not correspond to its Shapley value
contribution. The contributions instead depend on the feature distributions,
and on the subtle local interactions between the scoring features. ShaRP builds
on the Quantitative Input Influence framework, and can compute the
contributions of features for multiple Quantities of Interest, including score,
rank, pair-wise preference, and top-k. Because it relies on black-box access to
the ranker, ShaRP can be used to explain both score-based and learned ranking
models. We show results of an extensive experimental validation of ShaRP using
real and synthetic datasets, showcasing its usefulness for qualitative
analysis.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要