论文标题
深度神经网络加速器的硬件近似技术:一项调查
Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
论文作者
论文摘要
深度神经网络(DNN)非常受欢迎,因为它们在机器学习(ML)中的各种认知任务中具有高性能。在许多任务中,DNN的最新进展超出了人类的准确性,但以高计算复杂性为代价。为了有效地执行DNN推断,越来越多的研究工作利用了DNN的固有错误弹性并采用近似计算(AC)原理来解决DNN加速器的能源需求升高。本文提供了针对DNN加速器的硬件近似技术的全面调查和分析。首先,我们分析了艺术的状态并通过识别近似家族,我们将相对于近似类型的各个作品聚集。接下来,我们分析执行评估的复杂性(相对于数据集和DNN大小),以评估近似DNN加速器的效率,潜力和局限性。此外,提供了广泛的讨论,即错误指标更适合于设计DNN加速器的大约单元以及针对DNN推断量身定制的准确恢复方法。最后,我们介绍了DNN加速器的近似计算如何超越能效,并解决可靠性和安全问题。
Deep Neural Networks (DNNs) are very popular because of their high performance in various cognitive tasks in Machine Learning (ML). Recent advancements in DNNs have brought beyond human accuracy in many tasks, but at the cost of high computational complexity. To enable efficient execution of DNN inference, more and more research works, therefore, exploit the inherent error resilience of DNNs and employ Approximate Computing (AC) principles to address the elevated energy demands of DNN accelerators. This article provides a comprehensive survey and analysis of hardware approximation techniques for DNN accelerators. First, we analyze the state of the art and by identifying approximation families, we cluster the respective works with respect to the approximation type. Next, we analyze the complexity of the performed evaluations (with respect to the dataset and DNN size) to assess the efficiency, the potential, and limitations of approximate DNN accelerators. Moreover, a broad discussion is provided, regarding error metrics that are more suitable for designing approximate units for DNN accelerators as well as accuracy recovery approaches that are tailored to DNN inference. Finally, we present how Approximate Computing for DNN accelerators can go beyond energy efficiency and address reliability and security issues, as well.