A Domain-adaptive Pre-training Approach for Language Bias Detection in News

2022 ACM/IEEE Joint Conference on Digital Libraries (JCDL)(2022)

引用 7|浏览7
暂无评分
摘要
Media bias is a multi-faceted construct influencing individual behavior and collective decision-making. Slanted news reporting is the result of one-sided and polarized writing which can occur in various forms. In this work, we focus on an important form of media bias, i.e. bias by word choice. Detecting biased word choices is a challenging task due to its linguistic complexity and the lack of representative gold-standard corpora. We present DA-RoBERTa, a new state-of-the-art transformer-based model adapted to the media bias domain which identifies sentence-level bias with an F1 score of 0.814. In addition, we also train, DA-BERT and DA-BART, two more transformer models adapted to the bias domain. Our proposed domain-adapted models outperform prior bias detection approaches on the same data. CCS CONCEPTS • Computing methodologies → Natural language processing; • Information systems → Clustering and classification.
更多
查看译文
关键词
Media bias,news slant,neural classification,text analysis,domain adaptive
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要