Time estimation for deep learning model's inference in distributed processing units.

2023 IEEE Latin American Conference on Computational Intelligence (LA-CCI)(2023)

引用 0|浏览3
暂无评分
摘要
One problem with cloud computing is that it may fail to meet the desired time limits for real-time applications. In this regard, fog computing paradigm has gained ground as it complements the cloud by providing nodes with processing and storage capabilities closer to the data generation level. However, this level of the architecture has limited resources, making it necessary to efficiently distribute the workload involved in applications, especially when employing deep learning models. One technique to achieve this is task offloading, which involves distributing inference tasks throughout the architecture. Nevertheless, it is also important to know the time required for these tasks to be carried out within the network in order to obtain the desired response. In this work, we propose a queue-based convolutional neural network that allows estimating the response time for a deep learning inference task. Preliminary results demonstrate a good fit to the behavior of the datasets used in the experiment.
更多
查看译文
关键词
time estimation,convolution time series,fog computing,deep learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要