论文标题
从视频中蒸馏出管理法律和动态系统的来源输入
Distilling Governing Laws and Source Input for Dynamical Systems from Videos
论文作者
论文摘要
从视频中提取可解释的物理定律,由于深度学习的进步,最近对计算机视觉社区的兴趣扩大,但仍然是一个巨大的挑战。本文介绍了一个端到端的无监督深度学习框架,以根据记录的视频揭示移动对象提出的明确管理方程。取而代之的是,在图像空间的像素(空间)坐标系统中,物理定律是在回归的基础物理坐标系中建模的,物理状态遵循潜在的显式处理方程。设计了基于数值集成器的稀疏回归模块,并作为对自动编码器和坐标系回归的物理约束,同时,从学到的物理状态中揭示了简约的封闭形式的处理方程。模拟动态场景上的实验表明,所提出的方法能够提取封闭形式的管理方程,并同时识别视频记录的几种动态系统的未知激发输入,该系统填补了文献中没有现有方法的空白,并且可用于解决此类问题。
Distilling interpretable physical laws from videos has led to expanded interest in the computer vision community recently thanks to the advances in deep learning, but still remains a great challenge. This paper introduces an end-to-end unsupervised deep learning framework to uncover the explicit governing equations of dynamics presented by moving object(s), based on recorded videos. Instead in the pixel (spatial) coordinate system of image space, the physical law is modeled in a regressed underlying physical coordinate system where the physical states follow potential explicit governing equations. A numerical integrator-based sparse regression module is designed and serves as a physical constraint to the autoencoder and coordinate system regression, and, in the meanwhile, uncover the parsimonious closed-form governing equations from the learned physical states. Experiments on simulated dynamical scenes show that the proposed method is able to distill closed-form governing equations and simultaneously identify unknown excitation input for several dynamical systems recorded by videos, which fills in the gap in literature where no existing methods are available and applicable for solving this type of problem.