论文标题
部分可观测时空混沌系统的无模型预测
Active Importance Sampling for Variational Objectives Dominated by Rare Events: Consequences for Optimization and Generalization
论文作者
论文摘要
深度神经网络使用足够的数据进行优化时,提供了高维函数的准确表示;相比之下,在科学计算中占主导的功能近似技术并不能很好地扩展维度。结果,一旦被认为是棘手的高维抽样和近似问题,正在通过机器学习的镜头重新审视。虽然无与伦比的准确性的承诺可能暗示需要复兴的应用程序,该应用需要复杂系统的参数化表示形式,但在许多应用程序中收集足够数据以开发这种表示形式的应用程序仍然是一个重大挑战。在这里,我们介绍了一种将稀有事件采样技术与神经网络优化相结合的方法,以优化由罕见事件主导的目标函数。我们表明,重要性抽样可将解决方案的渐近差异减少到学习问题上,这表明了概括的益处。我们在学习系统的两个状态之间学习动态过渡途径的背景下研究我们的算法,这是统计物理学应用和机器学习理论中含义的应用问题。我们的数值实验表明,即使在高维度和稀有数据的复合困难中,我们也可以成功学习。
Deep neural networks, when optimized with sufficient data, provide accurate representations of high-dimensional functions; in contrast, function approximation techniques that have predominated in scientific computing do not scale well with dimensionality. As a result, many high-dimensional sampling and approximation problems once thought intractable are being revisited through the lens of machine learning. While the promise of unparalleled accuracy may suggest a renaissance for applications that require parameterizing representations of complex systems, in many applications gathering sufficient data to develop such a representation remains a significant challenge. Here we introduce an approach that combines rare events sampling techniques with neural network optimization to optimize objective functions that are dominated by rare events. We show that importance sampling reduces the asymptotic variance of the solution to a learning problem, suggesting benefits for generalization. We study our algorithm in the context of learning dynamical transition pathways between two states of a system, a problem with applications in statistical physics and implications in machine learning theory. Our numerical experiments demonstrate that we can successfully learn even with the compounding difficulties of high-dimension and rare data.