An Experimental Study of DNN Operator-Level Performance on Edge Devices

2023 IEEE International Conference on Smart Internet of Things (SmartIoT)(2023)

引用 0|浏览5
暂无评分
摘要
Deep Neural Network (DNN) has been widely used on edge devices with the development of Internet of Things (IoT). However, the limited resources of edge devices affect the deployment of DNN model on edge. Some works devote to deploying DNN models to edge devices have great demand on DNN inference latency prediction, but the typical prediction methods usually lack the knowledge of operator-level performance at DNN runtime, affecting the accuracy and generality. In this paper, we conduct a comprehensive experimental study of DNN operator-level performance on edge devices. We measured the inference latency performance of models and operator pairs in TensorRT and PyTorch frameworks on a Jetson Xavier NX Developer kit and a Jetson Nano Developer kit. Based on the experimental results, we summarize some useful hidden rules about inference latency relation between separated operators and operator blocks as well as operator fusion in different deep learning frameworks. Furthermore, we propose several implementation tips in aid of designing DNN inference latency predictor with higher accuracy and less complexity.
更多
查看译文
关键词
neural network,latency performance,edge device
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要