Interpret Neural Networks By Extracting Critical Subnetworks

IEEE TRANSACTIONS ON IMAGE PROCESSING(2020)

引用 8|浏览1934
暂无评分
摘要
In recent years, deep neural networks have achieved excellent performance in many fields of artificial intelligence. The requirements for the interpretability and robustness of neural networks are also increasing. In this paper, we propose to understand the functional mechanism of neural networks by extracting critical subnetworks. Specifically, we denote the critical subnetworks as a group of important channels across layers such that if they were suppressed to zeros, the final test performance would deteriorate severely. This novel perspective can not only reveal the layerwise semantic behavior within the model but also present more accurate visual explanations appearing in the data through attribution methods. Moreover, we propose two adversarial example detection methods based on the properties of sample-specific and class-specific subnetworks, which provides the possibility for increasing the model robustness.
更多
查看译文
关键词
Predictive models, Logic gates, Neural networks, Machine learning, Feature extraction, Robustness, Visualization, Model interpretability, model pruning, adversarial robustness
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要