论文标题
深层随机神经网络
Deep Randomized Neural Networks
论文作者
论文摘要
随机神经网络探讨了以随机或确定性方式固定大多数连接的神经系统的行为。此类系统的典型示例由多层神经网络体系结构组成,在初始化后,与隐藏层的连接未经训练。将训练算法限制在减少的权重固有的固有特征的一类具有许多有趣特征的随机神经网络的固有表征。其中,最终的学习过程的极端效率无疑是在训练有素的建筑方面的惊人优势。此外,尽管进行了简化,但在实践中,随机神经系统具有显着的特性,实现最新的属性会导致多个领域,从理论上讲,可以分析神经架构的内在特性(例如,在训练隐藏层连接之前)。近年来,对随机神经网络的研究已扩展到深度体系结构,为媒介和更复杂的数据域中的有效但非常有效的深度学习模型开辟了新的研究方向。本章调查有关随机神经网络的设计和分析的所有主要方面,以及有关其近似功能的一些关键结果。特别是,我们首先在馈送网络(即随机矢量功能链路和等效模型)和卷积过滤器的背景下首先介绍随机神经模型的基础,然后再转移到复发系统的情况(即储层计算网络)。对于这两者,我们都专门关注深度随机系统域的最新结果,以及(对于复发模型)它们在结构化域中的应用。
Randomized Neural Networks explore the behavior of neural systems where the majority of connections are fixed, either in a stochastic or a deterministic fashion. Typical examples of such systems consist of multi-layered neural network architectures where the connections to the hidden layer(s) are left untrained after initialization. Limiting the training algorithms to operate on a reduced set of weights inherently characterizes the class of Randomized Neural Networks with a number of intriguing features. Among them, the extreme efficiency of the resulting learning processes is undoubtedly a striking advantage with respect to fully trained architectures. Besides, despite the involved simplifications, randomized neural systems possess remarkable properties both in practice, achieving state-of-the-art results in multiple domains, and theoretically, allowing to analyze intrinsic properties of neural architectures (e.g. before training of the hidden layers' connections). In recent years, the study of Randomized Neural Networks has been extended towards deep architectures, opening new research directions to the design of effective yet extremely efficient deep learning models in vectorial as well as in more complex data domains. This chapter surveys all the major aspects regarding the design and analysis of Randomized Neural Networks, and some of the key results with respect to their approximation capabilities. In particular, we first introduce the fundamentals of randomized neural models in the context of feed-forward networks (i.e., Random Vector Functional Link and equivalent models) and convolutional filters, before moving to the case of recurrent systems (i.e., Reservoir Computing networks). For both, we focus specifically on recent results in the domain of deep randomized systems, and (for recurrent models) their application to structured domains.