论文标题
实现所有边缘深度学习:文学评论
Enabling All In-Edge Deep Learning: A Literature Review
论文作者
论文摘要
近年来,深度学习(DL)模型已经在诸如语音识别和自然语言理解等非平凡任务上取得了显着的成就。取得成功的重要因素之一是最终设备的扩散,该设备充当催化剂,以提供数据渴望数据的DL模型。但是,计算DL培训和推断是主要挑战。通常,中央云服务器用于计算,但它打开了其他重大挑战,例如高潜伏期,沟通成本增加和隐私问题。为了减轻这些缺点,已经做出了巨大的努力,以将DL模型的处理推向服务器。此外,DL和Edge的汇合点已引起到Edge Intelligence(EI)。该调查文件主要集中于EI的第五级,称为所有边缘水平,在该水平上,DL培训和推理(部署)仅由Edge服务器执行。当最终设备的计算资源低,例如,终端设备(例如,延迟和通信成本)在关键任务应用程序(例如医疗保健)中很重要时,所有边缘都是适合的。首先,本文介绍了所有边缘计算体系结构,包括集中式,分散和分发。其次,本文介绍了启用技术,例如模型并行性和分裂学习,这些技术促进了边缘服务器的DL培训和部署。第三,介绍了基于模型压缩和条件计算的模型适应技术,因为基于标准的基于云的DL部署由于其有限的计算资源而无法直接应用于所有边缘。第四,本文讨论了11个关键绩效指标,以有效地评估DL的性能。最后,提出了所有边缘领域的一些开放研究挑战。
In recent years, deep learning (DL) models have demonstrated remarkable achievements on non-trivial tasks such as speech recognition and natural language understanding. One of the significant contributors to its success is the proliferation of end devices that acted as a catalyst to provide data for data-hungry DL models. However, computing DL training and inference is the main challenge. Usually, central cloud servers are used for the computation, but it opens up other significant challenges, such as high latency, increased communication costs, and privacy concerns. To mitigate these drawbacks, considerable efforts have been made to push the processing of DL models to edge servers. Moreover, the confluence point of DL and edge has given rise to edge intelligence (EI). This survey paper focuses primarily on the fifth level of EI, called all in-edge level, where DL training and inference (deployment) are performed solely by edge servers. All in-edge is suitable when the end devices have low computing resources, e.g., Internet-of-Things, and other requirements such as latency and communication cost are important in mission-critical applications, e.g., health care. Firstly, this paper presents all in-edge computing architectures, including centralized, decentralized, and distributed. Secondly, this paper presents enabling technologies, such as model parallelism and split learning, which facilitate DL training and deployment at edge servers. Thirdly, model adaptation techniques based on model compression and conditional computation are described because the standard cloud-based DL deployment cannot be directly applied to all in-edge due to its limited computational resources. Fourthly, this paper discusses eleven key performance metrics to evaluate the performance of DL at all in-edge efficiently. Finally, several open research challenges in the area of all in-edge are presented.