Vid-ODE: Continuous-Time Video Generation
with Neural Ordinary Differential Equation

AAAI 2021

Jaegul Choo
KAIST
Joonseok Lee
Google Research
Sookyung Kim
Lawrence Livermore
Nat’l Lab.
Edword Choi
KAIST
Responsive image
Figure: Generating video frames in diverse time intervals based on a 5 FPS video. (Top row: Input to Vid-ODE. Remaining rows: Videos in various FPS between the start frame and the end frame.)

Abstract

Video generation models often operate under the assumption of fixed frame rates, which leads to suboptimal performance when it comes to handling flexible frame rates (e.g.,increasing the frame rate of more dynamic portion of the video as well as handling missing video frames). To resolve the restricted nature of existing video generation models’ ability to handle arbitrary timesteps, we propose continuous-time video generation by combining neural ODE (Vid-ODE) with pixel-level video processing techniques. Using ODE-ConvGRU as an encoder, a convolutional version of the recently proposed neural ODE, which enables us to learn continuous-time dynamics, Vid-ODE can learn the spatio-temporal dynamics of input videos of flexible frame rates. The decoder integrates the learned dynamics function to synthesize video frames at any given timesteps, where the pixel-level composition technique is used to maintain the sharpness of individual frames. With extensive experiments on four real-world video datasets, we verify that the proposed Vid-ODE outperforms state-of-the-art approaches under various video generation settings, both within the trained time range (interpolation) and beyond the range (extrapolation). To the best of our knowledge, Vid-ODE is the first work successfully performing continuous-time video generation using real-world videos.


Paper and Supplementary Material

[Paper] [Code]

AAAI, 2021.
Sunghyun Park*, Kangyeol Kim*, Junsoo Lee, Jaegul Choo, Joonseok Lee, Sookyung Kim, and Edward Choi. "Vid-ODE: Continuous-Time Video Generation with Neural Ordinary Differential Equation"


Method overview

Responsive image
Figure: Overview of Vid-ODE.

Additional Results

Responsive image
Figure: Qualitative comparisons with interpolation baselines on the Penn Action dataset.
Responsive image
Figure: Qualitative comparisons with extrapolation baselines on the Penn Action dataset.
Responsive image
Figure: Qualitative comparisons with interpolation baselines on the Penn Action dataset.
Responsive image
Figure: Generated video frames in diverse time intervals based on a 5-FPS input video with the Moving GIF dataset.