论文标题
Gennape:朝向广义的神经架构绩效估计器
GENNAPE: Towards Generalized Neural Architecture Performance Estimators
论文作者
论文摘要
预测神经体系结构的性能是一项具有挑战性的任务,对于神经建筑设计和搜索至关重要。现有方法要么依赖于涉及特定操作员和连接规则的预定义设计空间中建模架构的神经性能预测因子,而且不能推广到看不见的架构,或者诉诸于零成本的代理,这些代理并不总是准确。在本文中,我们提出了Gennape,这是一种广义的神经结构性能估计器,该估计器是在开放的神经体系结构基准上仔细预测的,旨在通过网络表示,对比预处理预处理和基于模糊的基于基于模糊的基于基于模糊的预测器的综合创新来概括完全看不见的体系结构。具体而言,Gennape代表一个给定的神经网络作为可以对任意体系结构进行建模的原子操作的计算图(CG)。它首先通过对比度学习学习了图形编码器,以鼓励通过拓扑特征分离网络,然后训练多个预测指标,根据神经网络的模糊成员资格,这些预测指标被软聚集。实验表明,在NAS-BENCH-101上预测的Gennape可以在没有或最低微调的情况下,在NAS-Bench-201,Nas-Bench-301,Mobilenet和Resnet家族中获得卓越的转移性,包括NAS-Bench-201,Nas-Bench-301,Mobilenet和Resnet家族。我们进一步介绍了3种挑战的新标记的神经网络基准:Hiaml,Inception和Twip Path,它们可以集中精度范围较窄。广泛的实验表明,Gennape可以正确辨别这些家庭中的高性能体系结构。最后,当与搜索算法配对时,Gennape可以找到提高准确性的架构,同时减少三个家庭的失败。
Predicting neural architecture performance is a challenging task and is crucial to neural architecture design and search. Existing approaches either rely on neural performance predictors which are limited to modeling architectures in a predefined design space involving specific sets of operators and connection rules, and cannot generalize to unseen architectures, or resort to zero-cost proxies which are not always accurate. In this paper, we propose GENNAPE, a Generalized Neural Architecture Performance Estimator, which is pretrained on open neural architecture benchmarks, and aims to generalize to completely unseen architectures through combined innovations in network representation, contrastive pretraining, and fuzzy clustering-based predictor ensemble. Specifically, GENNAPE represents a given neural network as a Computation Graph (CG) of atomic operations which can model an arbitrary architecture. It first learns a graph encoder via Contrastive Learning to encourage network separation by topological features, and then trains multiple predictor heads, which are soft-aggregated according to the fuzzy membership of a neural network. Experiments show that GENNAPE pretrained on NAS-Bench-101 can achieve superior transferability to 5 different public neural network benchmarks, including NAS-Bench-201, NAS-Bench-301, MobileNet and ResNet families under no or minimum fine-tuning. We further introduce 3 challenging newly labelled neural network benchmarks: HiAML, Inception and Two-Path, which can concentrate in narrow accuracy ranges. Extensive experiments show that GENNAPE can correctly discern high-performance architectures in these families. Finally, when paired with a search algorithm, GENNAPE can find architectures that improve accuracy while reducing FLOPs on three families.