CVPR 2021 Challenge on Super-Resolution Space

Overview

Learning the Super-Resolution Space Challenge
NTIRE 2021 at CVPR

Learning the Super-Resolution Space challenge is held as a part of the 6th edition of NTIRE: New Trends in Image Restoration and Enhancement workshop in conjunction with CVPR 2021. The goal of this challenge is to develop a super-resolution method that can actively sample from the space of plausible super-resolutions.

How to participate?

To participate in this challenge, please sign up using the following link and clone this repo to benchmark your results. Challenge participants can submit their paper to this CVPR 2021 Workshop.

CVPR 2021 Challenge Signup

Tackling the ill-posed nature of Super-Resolution

CVPR 2021 Challenge

Usually, super-resolution (SR) is trained using pairs of high- and low-resolution images. Infinitely many high-resolution images can be downsampled to the same low-resolution image. That means that the problem is ill-posed and cannot be inverted with a deterministic mapping. Instead, one can frame the SR problem as learning a stochastic mapping, capable of sampling from the space of plausible high-resolution images given a low-resolution image. This problem has been addressed in recent works [1, 2, 3]. The one-to-many stochastic formulation of the SR problem allows for a few potential advantages:

  • The development of more robust learning formulations that better accounts for the ill-posed nature of the SR problem.
  • Multiple predictions can be sampled and compared.
  • It opens the potential for controllable exploration and editing in the space of SR predictions.
Super-Resolution with Normalizing Flow Explorable SR Screenshot 2021-01-12 at 16 05 43
[Paper] [Project] [Paper] [Project] [Paper] [Project]
[1] SRFlow: Learning the Super-Resolution Space with Normalizing Flow. Lugmayr et al., ECCV 2020. [2] Explorable Super-Resolution. Bahat & Michaeli, CVPR 2020. [3] DeepSEE: Deep Disentangled Semantic Explorative Extreme Super-Resolution. Bühler et al., ACCV 2020.

CVPR 2021 Challenge on Learning the Super-Resolution Space

We organize this challenge to stimulate research in the emerging area of learning one-to-many SR mappings that are capable of sampling from the space of plausible solutions. Therefore the task is to develop a super-resolution method that:

  1. Each individual SR prediction should achieve highest possible photo-realism, as perceived by humans.
  2. Is capable of sampling an arbitrary number of SR images capturing meaningful diversity, corresponding to the uncertainty induced by the ill-posed nature of the SR problem together with image priors.
  3. Each individual SR prediction should be consistent with the input low-resolution image.

The challenge contains two tracks, targeting 4X and 8X super-resolution respectively. You can download the training and validation data in the table below. At a later stage, the low-resolution of the test set will be released.

  Training Validation
  Low-Resolution High-Resolution Low-Resolution High-Resolution
Track 4X 4X LR Train 4X HR Train 4X LR Valid 4X HR Valid
Track 8X 8X LR Train 8X HR Train 8X LR Valid 8X HR Valid

Challenge Rules

To guide the research towards useful and generalizable techniques, submissions need to adhere to the following rules. All participants must submit code of their solution along with the final results.

  • The method must be able to generate an arbitrary number of diverse samples. That is, your method cannot be limited to a maximum number of different SR samples (corresponding to e.g. a certain number of different output network heads).
  • All SR samples must be generated by a single model. That is, no ensembles are allowed.
  • No self-ensembles during inference (e.g. flipping and rotation).
  • All SR samples must be generated using the same hyper-parameters. That is, the generated SR samples shall not be the result of different choices of hyper-parameters during inference.
  • We accept submissions of deterministic methods. However, they will naturally score zero in the diversity measure and therefore not be able to win the challenge.
  • Other than the validation and test split of the DIV2k dataset, any training data or pre-training is allowed. You are not allowed to use DIV2K validation or test sets (low- and high-resolution images) for training.

Evaluation Protocol

A method is evaluated by first predicting a set of 10 randomly sampled SR images for each low-resolution image in the dataset. From this set of images, evaluation metrics corresponding to the three criteria above will be considered. The participating methods will be ranked according to each metric. These ranks will then be combined into a final score. The three evaluation metrics are described next.

git clone --recursive https://github.com/andreas128/NTIRE21_Learning_SR_Space.git
python3 measure.py OutName path/to/Ground-Truch path/to/Super-Resolution n_samples scale_factor

# n_samples = 10
# scale_factor = 4 for 4X and 8 for 8X

How we measure Photo-realism?

To assess the photo-realism, a human study will be performed on the test set for the final submission.

Automatically assessing the photo-realism and image quality is an extremely difficult task. All existing methods have severe shortcomings. As a very rough guide, you can use the LPIPS distance. Note: LPIPS will not be used to score photo-realism of you final submission. So beware of overfitting to LPIPS, as that can lead to worse results. LPIPS is integrated in our provided toolkit in measure.py.

How we measure the spanning of the SR Space?

The samples of the developed method should provide a meaningful diversity. To measure that, we define the following score. We sample 10 images, densely calculate a metric between the samples and the ground truth. To obtain the local best we pixel-wise select the best score out of the 10 samples and take the full image's average. The global best is obtained by averaging the whole image's score and selecting the best. Finally, we calculate the score using the following formula:

score = (global best - local best)/(global best) * 100

ESRGAN SRFlow
Track 4X 0 25.36
Track 8X 0 10.62

How we measure the Low Resolution Consistency

To measure how much information is preserved in the super-resloved image from the low-resolution image, we measure the LR-PSNR. The goal in this challenge is to obtain a LR-PSNR of 45dB. All approaches that have an average PSNR above this value will be ranked equally in terms of this criteria.

ESRGAN SRFlow
Track 4X 39.01 49.91
Track 8X 31.28 50.0

Important Dates

Date Event
2021.03.01 Final test data release (inputs only)
2021.03.08 test result submission deadline
2021.03.09 fact sheet / code / model submission deadline
2021.03.11 test preliminary score release to the participants
2021.03.28 challenge paper submission deadline
2021.04.13 camera-ready deadline
2021.06.15 workshop day

Submission of Final Test Results

After the final testing phase, participants will be asked to submit:

  • SR predictions on the test set.
  • Code.
  • A fact sheet describing their method.

Details will follow when the test phase starts ...

Issues and questions

In case of any questions about the challenge or the toolkit, feel free to open an issue on Github.

Organizers

CVPR 2021 NTIRE Terms and conditions

The terms and conditions for participating in the challenge are provided here

How to participate?

To participate in this challenge, please sign up using following link and clone this repo to benchmark your results. Challenge participants can submit their paper to this CVPR 2021 Workshop.

CVPR 2021 Challenge Signup

Owner
andreas
andreas
For IBM Quantum Challenge 2021 (May 20 - 26)

IBM Quantum Challenge 2021 Introduction Commemorating the 40-year anniversary of the Physics of Computation conference, and 5-year anniversary of IBM

Qiskit Community 140 Jan 01, 2023
Generalized Proximal Policy Optimization with Sample Reuse (GePPO)

Generalized Proximal Policy Optimization with Sample Reuse This repository is the official implementation of the reinforcement learning algorithm Gene

Jimmy Queeney 9 Nov 28, 2022
Matlab Python Heuristic Battery Opt - SMOP conversion and manual conversion

SMOP is Small Matlab and Octave to Python compiler. SMOP translates matlab to py

Tom Xu 1 Jan 12, 2022
Pytorch implementation of the DeepDream computer vision algorithm

deep-dream-in-pytorch Pytorch (https://github.com/pytorch/pytorch) implementation of the deep dream (https://en.wikipedia.org/wiki/DeepDream) computer

102 Dec 05, 2022
PyTorch and Tensorflow functional model definitions

functional-zoo Model definitions and pretrained weights for PyTorch and Tensorflow PyTorch, unlike lua torch, has autograd in it's core, so using modu

Sergey Zagoruyko 590 Dec 22, 2022
Learning Efficient Online 3D Bin Packing on Packing Configuration Trees

Learning Efficient Online 3D Bin Packing on Packing Configuration Trees This repository is being continuously updated, please stay tuned! Any code con

86 Dec 28, 2022
Impelmentation for paper Feature Generation and Hypothesis Verification for Reliable Face Anti-Spoofing

FGHV Impelmentation for paper Feature Generation and Hypothesis Verification for Reliable Face Anti-Spoofing Requirements Python 3.6 Pytorch 1.5.0 Cud

5 Jun 02, 2022
Hierarchical Metadata-Aware Document Categorization under Weak Supervision (WSDM'21)

Hierarchical Metadata-Aware Document Categorization under Weak Supervision This project provides a weakly supervised framework for hierarchical metada

Yu Zhang 53 Sep 17, 2022
The code for our paper submitted to RAL/IROS 2022: OverlapTransformer: An Efficient and Rotation-Invariant Transformer Network for LiDAR-Based Place Recognition.

OverlapTransformer The code for our paper submitted to RAL/IROS 2022: OverlapTransformer: An Efficient and Rotation-Invariant Transformer Network for

HAOMO.AI 136 Jan 03, 2023
FlowTorch is a PyTorch library for learning and sampling from complex probability distributions using a class of methods called Normalizing Flows

FlowTorch is a PyTorch library for learning and sampling from complex probability distributions using a class of methods called Normalizing Flows.

Meta Incubator 272 Jan 02, 2023
Code for "Diversity can be Transferred: Output Diversification for White- and Black-box Attacks"

Output Diversified Sampling (ODS) This is the github repository for the NeurIPS 2020 paper "Diversity can be Transferred: Output Diversification for W

50 Dec 11, 2022
Implementation of Self-supervised Graph-level Representation Learning with Local and Global Structure (ICML 2021).

Self-supervised Graph-level Representation Learning with Local and Global Structure Introduction This project is an implementation of ``Self-supervise

MilaGraph 50 Dec 09, 2022
[NeurIPS 2021] Source code for the paper "Qu-ANTI-zation: Exploiting Neural Network Quantization for Achieving Adversarial Outcomes"

Qu-ANTI-zation This repository contains the code for reproducing the results of our paper: Qu-ANTI-zation: Exploiting Quantization Artifacts for Achie

Secure AI Systems Lab 8 Mar 26, 2022
Very simple NCHW and NHWC conversion tool for ONNX. Change to the specified input order for each and every input OP. Also, change the channel order of RGB and BGR. Simple Channel Converter for ONNX.

scc4onnx Very simple NCHW and NHWC conversion tool for ONNX. Change to the specified input order for each and every input OP. Also, change the channel

Katsuya Hyodo 16 Dec 22, 2022
A Rao-Blackwellized Particle Filter for 6D Object Pose Tracking

PoseRBPF: A Rao-Blackwellized Particle Filter for 6D Object Pose Tracking PoseRBPF Paper Self-supervision Paper Pose Estimation Video Robot Manipulati

NVIDIA Research Projects 107 Dec 25, 2022
A Python library for unevenly-spaced time series analysis

traces A Python library for unevenly-spaced time series analysis. Why? Taking measurements at irregular intervals is common, but most tools are primar

Datascope Analytics 516 Dec 29, 2022
PyTorch implementation of "Contrast to Divide: self-supervised pre-training for learning with noisy labels"

Contrast to Divide: self-supervised pre-training for learning with noisy labels This is an official implementation of "Contrast to Divide: self-superv

55 Nov 23, 2022
Keras implementation of the GNM model in paper ’Graph-Based Semi-Supervised Learning with Nonignorable Nonresponses‘

Graph-based joint model with Nonignorable Missingness (GNM) This is a Keras implementation of the GNM model in paper ’Graph-Based Semi-Supervised Lear

Fan Zhou 2 Apr 17, 2022
A New Open-Source Off-road Environment for Benchmark Generalization of Autonomous Driving

A New Open-Source Off-road Environment for Benchmark Generalization of Autonomous Driving Isaac Han, Dong-Hyeok Park, and Kyung-Joong Kim IEEE Access

13 Dec 27, 2022
HHP-Net: A light Heteroscedastic neural network for Head Pose estimation with uncertainty

HHP-Net: A light Heteroscedastic neural network for Head Pose estimation with uncertainty Giorgio Cantarini, Francesca Odone, Nicoletta Noceti, Federi

18 Aug 02, 2022