Impact of Tensor Cores and Mixed Precision on the Reliability of Matrix Multiplication in GPUs

IEEE Transactions on Nuclear Science(2020)

引用 11|浏览5
暂无评分
摘要
Matrix multiplication (MxM) is a cornerstone application for both high-performance computing and safety-critical applications. Most of the operations in convolutional neural networks for object detection, in fact, are MxM related. Chip designers are proposing novel solutions to improve the efficiency of the execution of MxM. In this article, we investigate the impact of two novel architectures for MxM (i.e., tensor cores and mixed precision) on the graphics processing units (GPUs) reliability. In addition, we evaluate how effective the embedded error-correcting code is in reducing the MxM error rate. Our results show that low-precision operations are more reliable, and the tensor core increases the amount of data correctly produced by the GPU. However, reducing precision and the use of tensor core significantly increase the impact of faults in the output correctness.
更多
查看译文
关键词
Graphics processing unit (GPU),matrix multiplication (MxM),neutrons,reliability,soft errors,tensor core
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要