论文标题

通过舞台的神经建筑搜索

Stage-Wise Neural Architecture Search

论文作者

Jordao, Artur, Akio, Fernando, Lie, Maiko, Schwartz, William Robson

论文摘要

Resnet和Nasnet等现代卷积网络已在许多计算机视觉应用中实现了最先进的结果。这些体系结构由阶段组成,这些阶段是在同一分辨率中的表示形式下运行的一组。已经证明,增加每个阶段的层数可以提高网络的预测能力。但是,从浮点操作,内存需求和推理时间方面,所得的体系结构在计算上变得昂贵。因此,为评估深度和绩效之间的不同权衡是必要的。为了解决这个问题,最近的作品提议自动设计高性能体系结构,主要是通过神经体系结构搜索(NAS)。当前的NAS策略分析了大量可能的候选架构,因此需要大量的计算资源,并花费了许多GPU天。在此激励的情况下,我们提出了一种NAS方法,以有效设计准确和低成本的卷积体系结构,并证明设计这些体系结构的有效策略是学习深度阶段。为此,考虑到其重要性,我们的方法在每个阶段都会逐步提高深度,因此,重要性较低的阶段在较低的阶段变得更深。我们对CIFAR和不同版本的Imagenet数据集进行了实验,在这里我们表明,通过我们的方法发现的体系结构比人为设计的体系结构具有更好的准确性和效率。此外,我们表明在CIFAR-10上发现的体系结构可以成功传输到大型数据集。与以前的NAS方法相比,我们的方法要高得多,因为它评估了一个数量级的模型,并且与最先进的架构相当。

Modern convolutional networks such as ResNet and NASNet have achieved state-of-the-art results in many computer vision applications. These architectures consist of stages, which are sets of layers that operate on representations in the same resolution. It has been demonstrated that increasing the number of layers in each stage improves the prediction ability of the network. However, the resulting architecture becomes computationally expensive in terms of floating point operations, memory requirements and inference time. Thus, significant human effort is necessary to evaluate different trade-offs between depth and performance. To handle this problem, recent works have proposed to automatically design high-performance architectures, mainly by means of neural architecture search (NAS). Current NAS strategies analyze a large set of possible candidate architectures and, hence, require vast computational resources and take many GPUs days. Motivated by this, we propose a NAS approach to efficiently design accurate and low-cost convolutional architectures and demonstrate that an efficient strategy for designing these architectures is to learn the depth stage-by-stage. For this purpose, our approach increases depth incrementally in each stage taking into account its importance, such that stages with low importance are kept shallow while stages with high importance become deeper. We conduct experiments on the CIFAR and different versions of ImageNet datasets, where we show that architectures discovered by our approach achieve better accuracy and efficiency than human-designed architectures. Additionally, we show that architectures discovered on CIFAR-10 can be successfully transferred to large datasets. Compared to previous NAS approaches, our method is substantially more efficient, as it evaluates one order of magnitude fewer models and yields architectures on par with the state-of-the-art.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源