Abstract
Learning workable representations of dynamical systems is becoming an
increasingly important problem in a number of application areas. By leveraging
recent work connecting deep neural networks to systems of differential
equations, we propose variational integrator networks, a class of neural
network architectures designed to ensure faithful representations of the
dynamics under study. This class of network architectures facilitates accurate
long-term prediction, interpretability, and data-efficient learning, while
still remaining highly flexible and capable of modeling complex behavior. We
demonstrate that they can accurately learn dynamical systems from both noisy
observations in phase space and from image pixels within which the unknown
dynamics are embedded.
Users
Please
log in to take part in the discussion (add own reviews or comments).