Generating Realistic Videos From Keyframes With Concatenated GANs

Publisher:
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
Publication Type:
Journal Article
Citation:
IEEE Transactions on Circuits and Systems for Video Technology, 2019, 29, (8), pp. 2337-2348
Issue Date:
2019-08-01
Filename Description Size
Generating_Realistic_Videos_From_Keyframes_With_Concatenated_GANs.pdfPublished version3.98 MB
Adobe PDF
Full metadata record
Given two video frames X0 and Xn+1, we aim to generate a series of intermediate frames Y1, Y2, . . ., Yn, such that the resulting video consisting of frames X0, Y1 − Yn, andXn+1 appears realistic to a human watcher. Such video generation has numerous important applications, including video compression, movie production, slow-motion filming, video surveillance, and forensic analysis. Yet, video generation is highly challenging due to the vast search space of possible frames. Previous methods, mostly based on video prediction and/or video interpolation, tend to generate poor-quality videos with severe motion blur. This paper proposes a novel, end-to-end approach to video generation using generative adversarial networks (GANs). In particular, our design involves two concatenated GANs, one capturing motions and the other generating frame details. The loss function is also carefully engineered to include adversarial loss, gradient difference (for motion learning), and normalized product correlation loss (for frame details). Experiments using three video datasets, namely, Google Robotic Push, KTH human actions, and UCF101, demonstrate that the proposed solution generates high-quality, realistic, and sharp videos, whereas all previous solutions output noisy and blurry results.
Please use this identifier to cite or link to this item: