Poisoning Attacks against Recommender Systems: A Survey
CoRR(2024)
摘要
Modern recommender systems (RS) have seen substantial success, yet they
remain vulnerable to malicious activities, notably poisoning attacks. These
attacks involve injecting malicious data into the training datasets of RS,
thereby compromising their integrity and manipulating recommendation outcomes
for gaining illicit profits. This survey paper provides a systematic and
up-to-date review of the research landscape on Poisoning Attacks against
Recommendation (PAR). A novel and comprehensive taxonomy is proposed,
categorizing existing PAR methodologies into three distinct categories:
Component-Specific, Goal-Driven, and Capability Probing. For each category, we
discuss its mechanism in detail, along with associated methods. Furthermore,
this paper highlights potential future research avenues in this domain.
Additionally, to facilitate and benchmark the empirical comparison of PAR, we
introduce an open-source library, ARLib, which encompasses a comprehensive
collection of PAR models and common datasets. The library is released at
https://github.com/CoderWZW/ARLib.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要