Superresolution
Super-Resolution Erlangen (SupER)
Benchmarking Super-Resolution Algorithms on Real Data
Overview
Capturing ground truth data to benchmark super-resolution (SR) is challenging. Therefore, current quantitative studies are mainly evaluated on simulated data artificially sampled from ground truth images. We argue that such evaluations overestimate the actual performance of SR methods compared to their behavior on real images.
Toward bridging this simulated-to-real gap, we introduce the Super-Resolution Erlangen (SupER) database, the first comprehensive laboratory SR database of all-real acquisitions with pixel-wise ground truth. It consists of more than 80k images of 14 scenes combining different facets: CMOS sensor noise, real sampling at four resolution levels, nine scene motion types, two photometric conditions, and lossy video coding at five levels. As such, the database exceeds existing benchmarks by an order of magnitude in quality and quantity. This work also benchmarks 19 popular single-image and multi-frame algorithms on our data. The benchmark comprises a quantitative study by exploiting ground truth data and qualitative evaluations in a large-scale observer study. We also rigorously investigate agreements between both evaluations from a statistical perspective. One interesting result is that top-performing methods on simulated data may be surpassed by others on real data. Our insights can spur further algorithm development, and the publicy available dataset can foster future evaluations.
Database
The images were captured with a monochromatic Basler acA2000-50gm CMOS camera to avoid subsampling introduced by a Bayer pattern and are stored in PNG format (8 bit, grayscale). Each sequence has 40 frames and is available in 4 spatial resolution levels using hardware binning: original (2040×1080), 2×2 binning (1020×540), 3×3 binning (680×360), and 4×4 binning (510×270). The sequences cover multiple types of camera and object motion. Furthermore, besides the regular sequences (inliers), we also provide the same sequences with photometric outliers (5 of 40 frames were captured with significantly less light in the scene) and video compression using H.265/HEVC coding (4 compression levels).
The sequence for 14 scenes including all motion types, all resolution levels, all compression levels, and both inliers and outliers can be downloaded by clicking on the respective sequence name (~2.5 GB per sequence).
| Banknotes | Books-and-papers | Bookshelf | Coffee | Dolls |
| Games | Globe | Globe-fast | Loader | Newspapers |
| Pencils | Porsche | Tea-bottles | Truck-budha-duck |
Source Code
Our source code including all evaluation protocols and implementations of the benchmarked SR algorithms is available at github or here.
Benchmark Results
Quantitative Study
Here, we provide the results of our quantitative evaluations including SR images and image quality measures. The results can be downloaded separately for the different algorithms. Each algorithm can be identified by its ID. The ground truth data is accessible via ID = 0. Please see the evaluation framework for details on how to analyze these results.
Human Observer Study
The results of our human observer study can be downloaded here.
Time Measurements
Measurements regarding the computation times of the different algorithms can be downloaded here.
Publication
If you use the dataset or source code for your research, you should cite the follwing paper:
Thomas Köhler Michel Bätz, Farzad Naderi, André Kaup, Andreas Maier, Christian Riess
Toward Bridging the Simulated-to-Real Gap: Benchmarking Super-Resolution on Real Data
IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, num. 11, Nov., 2020, pp. 2944-2959
DOI: https://dx.doi.org/10.1109/TPAMI.2019.2917037
