scholarly journals Channel Attention Is All You Need for Video Frame Interpolation

2020 ◽  
Vol 34 (07) ◽  
pp. 10663-10671 ◽  
Author(s):  
Myungsub Choi ◽  
Heewon Kim ◽  
Bohyung Han ◽  
Ning Xu ◽  
Kyoung Mu Lee

Prevailing video frame interpolation techniques rely heavily on optical flow estimation and require additional model complexity and computational cost; it is also susceptible to error propagation in challenging scenarios with large motion and heavy occlusion. To alleviate the limitation, we propose a simple but effective deep neural network for video frame interpolation, which is end-to-end trainable and is free from a motion estimation network component. Our algorithm employs a special feature reshaping operation, referred to as PixelShuffle, with a channel attention, which replaces the optical flow computation module. The main idea behind the design is to distribute the information in a feature map into multiple channels and extract motion information by attending the channels for pixel-level frame synthesis. The model given by this principle turns out to be effective in the presence of challenging motion and occlusion. We construct a comprehensive evaluation benchmark and demonstrate that the proposed approach achieves outstanding performance compared to the existing models with a component for optical flow computation.

2020 ◽  
Vol 224 ◽  
pp. 01027
Author(s):  
P. V. Belyakov ◽  
M. B. Nikiforov ◽  
E. R. Muratov ◽  
O. V. Melnik

Optical flow computation is one of the most important tasks in computer vision. The article deals with a modification of the variational method of the optical flow computation, according to its application in stereo vision. Such approaches are traditionally based on a brightness constancy assumption and a gradient constancy assumption during pixels motion. Smoothness assumption also restricts motion discontinuities, i.e. the smoothness of the vector field of pixel velocity is assumed. It is proposed to extend the functional of the optical flow computation in a similar way by adding a priori known stereo cameras extrinsic parameters and minimize such jointed model of optical flow computation. The article presents a partial differential equations framework in image processing and numerical scheme for its implementation. Performed experimental evaluation demonstrates that the proposed method gives smaller errors than traditional methods of optical flow computation.


2014 ◽  
Vol 556-562 ◽  
pp. 4352-4356
Author(s):  
Jun Wu ◽  
Ming Cheng Luo ◽  
Jun Li

UAV Video is rapidly emerging as a widely used source of imagery for many applications in recent years. This paper presents our research on the mosaic of UAV video for the purpose of harbor surveillance. First, one new framework on estimating video frame transformation with Optical flow is presented in this paper. For this new framework, fewer number of Gaussian pyramid is created for implementing for the multiresolution approach and thus, more details for optical flow computation is well kept; Second, we make a discussion on using Fourier-Mellin Transformation in image frequency domain to estimate initial motion parameter of adjacent video frames and with those initial motion parameters, small displacements for optical flow computation can be achieved; The experimental results demonstrated that the mosaic image generated from aerial video shows satisfied visual quality and its surveillance application for fast response to time-critical event, e.g., flood, is descried.


Author(s):  
Minseop Kim ◽  
Haechul Choi

Recently, the demand for high-quality video content has rapidly been increasing, led by the development of network technology and the growth in video streaming platforms. In particular, displays with a high refresh rate, such as 120 Hz, have become popular. However, the visual quality is only enhanced if the video stream is produced at the same high frame rate. For the high quality, conventional videos with a low frame rate should be converted into a high frame rate in real time. This paper introduces a bidirectional intermediate flow estimation method for real-time video frame interpolation. A bidirectional intermediate optical flow is directly estimated to predict an accurate intermediate frame. For real-time processing, multiple frames are interpolated with a single intermediate optical flow and parts of the network are implemented in 16-bit floating-point precision. Perceptual loss is also applied to improve the cognitive performance of the interpolated frames. The experimental results showed a high prediction accuracy of 35.54 dB on the Vimeo90K triplet benchmark dataset. The interpolation speed of 84 fps was achieved for 480p resolution.


Sign in / Sign up

Export Citation Format

Share Document