Verification of integrity of deployed deep learning models using Bayesian Optimization

Knowledge-Based Systems(2022)

引用 5|浏览1
暂无评分
摘要
Artificial Intelligence (AI) is hugely driven by machine learning (ML). Popular ML algorithms like deep neural networks (DNN) are used in wide variety of applications. Due to excellent predictive capabilities, these models are used in many realworld applications. To accommodate the huge service demands from these models, they are hosted on cloud platforms. Hosting a model in cloud raises security concerns. Malicious third party can modify the model when in transit from local system to cloud platform or after being uploaded in the cloud. We propose Bayesian Compromise Detection (BCD) algorithm to detect such model compromises by generating sensitive samples. Finding a sensitive sample is essentially an optimization problem that aims to maximize the prediction differences between the original and the compromised model. The optimization problem is challenging as (1) a cloud customer can only have a blackbox access to the compromised model (2) sensitive sample has to be searched in the distribution of the training inputs, which is most likely a high-dimensional space and (3) the optimization problem is non-convex. To tackle these obstacles, we use Variational Autoencoder (VAE) to map high-dimensional data to a non-linear low-dimensional space and then use Bayesian Optimization (BO) to find the globally optimum sensitive sample. Proposed algorithm generates a sensitive sample that can detect model compromise without incurring much cost by multiple queries. Experiments with multiple datasets demonstrate the effectiveness of our proposed method. Our method outperforms the state-of-the-art method in terms of detection rate of the compromised models on cloud platforms.
更多
查看译文
关键词
Cloud service,Sensitive sample,Bayesian optimization,Trojan model
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要