A single video super-resolution GAN for multiple downsampling operators based on pseudo-inverse image formation models
Abstract
The popularity of high and ultra-high definition displays has led to the need for methods to improve the quality of videos already obtained at much lower resolutions. A large amount of current CNN-based Video Super-Resolution methods are designed and trained to handle a specific degradation operator (e.g., bicubic downsampling) and are not robust to mismatch between training and testing degradation models. This causes their performance to deteriorate in real-life applications. Furthermore, many of them use the Mean-Squared-Error as the only loss during learning, causing the resulting images to be too smooth. In this work we propose a new Convolutional Neural Network for video super resolution which is robust to multiple degradation models. During training, which is performed on a large dataset of scenes with slow and fast motions, it uses the pseudo-inverse image formation model as part of the network architecture in conjunction with perceptual losses and a smoothness constraint that eliminates the artifacts originating from these perceptual losses. The experimental validation shows that our approach outperforms current state-of-the-art methods and is robust to multiple degradations.
- Publication:
-
Digital Signal Processing
- Pub Date:
- September 2020
- DOI:
- arXiv:
- arXiv:1907.01399
- Bibcode:
- 2020DSP...10402801L
- Keywords:
-
- Video;
- Super-resolution;
- Convolutional neuronal networks;
- Generative adversarial networks;
- Perceptual loss functions;
- Computer Science - Computer Vision and Pattern Recognition;
- Computer Science - Machine Learning;
- Statistics - Machine Learning
- E-Print:
- doi:10.1016/j.dsp.2020.102801