订阅小程序
旧版功能

Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs

Aodong Chen,Fei Xu,Li Han, Yuan Dong,Li Chen,Zhi Zhou,Fangming Liu

IEEE TRANSACTIONS ON COMPUTERS(2025)

引用 0|浏览21
关键词
Graphics processing units,Streams,Parallel processing,Artificial neural networks,Solid modeling,Resource management,Kernel,Interference,Computational modeling,Runtime,DNN inference,DNN operator parallelism,scheduling,GPU resource utilization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要