论文标题
紧凑的深神经网络有效合成
Efficient Synthesis of Compact Deep Neural Networks
论文作者
论文摘要
深度神经网络(DNN)已被部署在众多机器学习应用中。但是,通过日益复杂和深度的网络体系结构,通常可以取得进步的准确性。这些大型的深层模型通常不适合实际应用,因为它们的计算成本很高,记忆带宽和延迟延迟。例如,自主驾驶需要基于在运行时能量和内存存储限制下运行的电网(IoT)边缘设备进行快速推断。在这种情况下,紧凑的DNN可以促进部署,因为它们的能耗减少,记忆需求和推理潜伏期。长期的短期记忆(LSTMS)是一种复发性神经网络,在顺序数据建模的背景下也发现了广泛使用。他们还面临模型大小与准确性权衡。在本文中,我们回顾了自动合成紧凑型但准确的DNN/LSTM模型的主要方法。我们还概述了一些挑战和未来的探索领域。
Deep neural networks (DNNs) have been deployed in myriad machine learning applications. However, advances in their accuracy are often achieved with increasingly complex and deep network architectures. These large, deep models are often unsuitable for real-world applications, due to their massive computational cost, high memory bandwidth, and long latency. For example, autonomous driving requires fast inference based on Internet-of-Things (IoT) edge devices operating under run-time energy and memory storage constraints. In such cases, compact DNNs can facilitate deployment due to their reduced energy consumption, memory requirement, and inference latency. Long short-term memories (LSTMs) are a type of recurrent neural network that have also found widespread use in the context of sequential data modeling. They also face a model size vs. accuracy trade-off. In this paper, we review major approaches for automatically synthesizing compact, yet accurate, DNN/LSTM models suitable for real-world applications. We also outline some challenges and future areas of exploration.