A curated (most recent) list of resources for Learning with Noisy Labels

Overview

Learning-with-Noisy-Labels

A curated list of most recent papers & codes in Learning with Noisy Labels


Papers & Code in 2021

This repo focus on papers after 2019, for previous works, please refer to (https://github.com/subeeshvasu/Awesome-Learning-with-Label-Noise).

ICML 2021

Conference date: Jul 18, 2021 -- Jul 24, 2021

  • [UCSC REAL Lab] The importance of understanding instance-level noisy labels. [Paper]
  • [UCSC REAL Lab] Clusterability as an Alternative to Anchor Points When Learning with Noisy Labels. [Paper][Code]
  • Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision. [Paper][Code]
  • Learning Noise Transition Matrix from Only Noisy Labels via Total Variation Regularization. [Paper][Code]
  • Class2Simi: A Noise Reduction Perspective on Learning with Noisy Labels. [Paper]
  • Provably End-to-end Label-noise Learning without Anchor Points. [Paper]
  • Asymmetric Loss Functions for Learning with Noisy Labels. [Paper][Code]
  • Confidence Scores Make Instance-dependent Label-noise Learning Possible. [Paper]
  • Provable Generalization of SGD-trained Neural Networks of Any Width in the Presence of Adversarial Label Noise. [Paper]
  • Wasserstein Distributional Normalization For Robust Distributional Certification of Noisy Labeled Data. [Paper]
  • Learning from Noisy Labels with No Change to the Training Process. [Paper]

ICLR 2021

  • [UCSC REAL Lab] When Optimizing f-Divergence is Robust with Label Noise. [Paper][Code]
  • [UCSC REAL Lab] Learning with Instance-Dependent Label Noise: A Sample Sieve Approach. [Paper][Code]
  • Noise against noise: stochastic label noise helps combat inherent label noise. [Paper][Code]
  • Learning with Feature-Dependent Label Noise: A Progressive Approach. [Paper][Code]
  • Robust early-learning: Hindering the memorization of noisy labels. [Paper][Code]
  • MoPro: Webly Supervised Learning with Momentum Prototypes. [Paper] [Code]
  • Robust Curriculum Learning: from clean label detection to noisy label self-correction. [Paper]
  • How Does Mixup Help With Robustness and Generalization? [Paper]
  • Theoretical Analysis of Self-Training with Deep Networks on Unlabeled Data. [Paper]

CVPR 2021

Conference date: Jun 19, 2021 -- Jun 25, 2021

  • [UCSC REAL Lab] A Second-Order Approach to Learning with Instance-Dependent Label Noise. [Paper][Code]
  • Improving Unsupervised Image Clustering With Robust Learning. [Paper]
  • Multi-Objective Interpolation Training for Robustness to Label Noise. [Paper][Code]
  • Noise-resistant Deep Metric Learning with Ranking-based Instance Selection. [Paper][Code]
  • Augmentation Strategies for Learning with Noisy Labels. [Paper][Code]
  • Jo-SRC: A Contrastive Approach for Combating Noisy Labels. [Paper][Code]
  • Multi-Objective Interpolation Training for Robustness to Label Noise. [Paper][Code]
  • Partially View-aligned Representation Learning with Noise-robust Contrastive Loss. [Paper][Code]
  • Correlated Input-Dependent Label Noise in Large-Scale Image Classification. [Paper]
  • DAT: Training Deep Networks Robust To Label-Noise by Matching the Feature Distributions.[Paper]
  • Faster Meta Update Strategy for Noise-Robust Deep Learning. [Paper][Code]
  • DualGraph: A graph-based method for reasoning about label noise. [Paper]
  • Background-Aware Pooling and Noise-Aware Loss for Weakly-Supervised Semantic Segmentation. [Paper]
  • Joint Negative and Positive Learning for Noisy Labels. [Paper]
  • Faster Meta Update Strategy for Noise-Robust Deep Learning. [Paper]
  • AutoDO: Robust AutoAugment for Biased Data with Label Noise via Scalable Probabilistic Implicit Differentiation. [Paper][Code]
  • Meta Pseudo Labels. [Paper][Code]
  • All Labels Are Not Created Equal: Enhancing Semi-supervision via Label Grouping and Co-training. [Paper][Code]
  • SimPLE: Similar Pseudo Label Exploitation for Semi-Supervised Classification. [Paper][Code]

AISTATS 2021

Conference date: Apr 13, 2021 -- Apr 15, 2021

  • Collaborative Classification from Noisy Labels. [Paper]
  • Linear Models are Robust Optimal Under Strategic Behavior. [Paper]

AAAI 2021

  • Beyond Class-Conditional Assumption: A Primary Attempt to Combat Instance-Dependent Label Noise. [Paper][Code]
  • Learning to Purify Noisy Labels via Meta Soft Label Corrector. [Paper][Code]
  • Robustness of Accuracy Metric and its Inspirations in Learning with Noisy Labels. [Paper][Code]
  • Learning from Noisy Labels with Complementary Loss Functions. [Paper][Code]
  • Analysing the Noise Model Error for Realistic Noisy Label Data. [Paper][Code]
  • Tackling Instance-Dependent Label Noise via a Universal Probabilistic Model. [Paper]
  • Learning with Group Noise. [Paper]
  • Meta Label Correction for Noisy Label Learning. [Paper]

ArXiv 2021

  • [UCSC REAL Lab] Understanding (Generalized) Label Smoothing when Learning with Noisy Labels. [Paper]
  • Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks. [Paper][Code]
  • Estimating Instance-dependent Label-noise Transition Matrix using DNNs. [Paper]
  • A Theoretical Analysis of Learning with Noisily Labeled Data. [Paper]
  • Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels. [Paper]
  • A Survey of Label-noise Representation Learning: Past, Present and Future. [Paper]
  • Learning Noise Transition Matrix from Only Noisy Labels via Total Variation Regularization. [Paper][Code]
  • Noisy-Labeled NER with Confidence Estimation. [Paper][Code]
  • Study Group Learning: Improving Retinal Vessel Segmentation Trained with Noisy Labels. [Paper][Code]
  • Contrast to Divide: Self-Supervised Pre-Training for Learning with Noisy Labels. [Paper][Code]
  • Exponentiated Gradient Reweighting for Robust Training Under Label Noise and Beyond. [Paper]
  • Understanding the Interaction of Adversarial Training with Noisy Labels. [Paper]
  • Learning from Noisy Labels via Dynamic Loss Thresholding. [Paper]
  • Evaluating Multi-label Classifiers with Noisy Labels. [Paper]
  • Self-Supervised Noisy Label Learning for Source-Free Unsupervised Domain Adaptation. [Paper]
  • Transform consistency for learning with noisy labels. [Paper]
  • Learning to Combat Noisy Labels via Classification Margins. [Paper]
  • Joint Negative and Positive Learning for Noisy Labels. [Paper]
  • Robust Classification from Noisy Labels: Integrating Additional Knowledge for Chest Radiography Abnormality Assessment. [Paper]
  • DST: Data Selection and joint Training for Learning with Noisy Labels. [Paper]
  • LongReMix: Robust Learning with High Confidence Samples in a Noisy Label Environment. [Paper]
  • A Novel Perspective for Positive-Unlabeled Learning via Noisy Labels. [Paper]
  • Ensemble Learning with Manifold-Based Data Splitting for Noisy Label Correction. [Paper]
  • MetaLabelNet: Learning to Generate Soft-Labels from Noisy-Labels. [Paper]
  • On the Robustness of Monte Carlo Dropout Trained with Noisy Labels. [Paper]
  • Co-matching: Combating Noisy Labels by Augmentation Anchoring. [Paper]
  • Pathological Image Segmentation with Noisy Labels. [Paper]
  • CrowdTeacher: Robust Co-teaching with Noisy Answers & Sample-specific Perturbations for Tabular Data. [Paper]
  • Approximating Instance-Dependent Noise via Instance-Confidence Embedding. [Paper]
  • Rethinking Noisy Label Models: Labeler-Dependent Noise with Adversarial Awareness. [Paper]
  • ScanMix: Learning from Severe Label Noise viaSemantic Clustering and Semi-Supervised Learning. [Paper]
  • Friends and Foes in Learning from Noisy Labels. [Paper]
  • Learning from Noisy Labels for Entity-Centric Information Extraction. [Paper]
  • A Fremework Using Contrastive Learning for Classification with Noisy Labels. [Paper]
  • Contrastive Learning Improves Model Robustness Under Label Noise. [Paper][Code]
  • Noise-Resistant Deep Metric Learning with Probabilistic Instance Filtering. [Paper]
  • Compensation Learning. [Paper]
  • kNet: A Deep kNN Network To Handle Label Noise. [Paper]
  • Temporal-aware Language Representation Learning From Crowdsourced Labels. [Paper]
  • Memorization in Deep Neural Networks: Does the Loss Function matter?. [Paper]
  • Mitigating Memorization in Sample Selection for Learning with Noisy Labels. [Paper]
  • P-DIFF: Learning Classifier with Noisy Labels based on Probability Difference Distributions. [Paper][Code]
  • Decoupling Representation and Classifier for Noisy Label Learning. [Paper]
  • Contrastive Representations for Label Noise Require Fine-Tuning. [Paper]
  • NGC: A Unified Framework for Learning with Open-World Noisy Data. [Paper]
  • Learning From Long-Tailed Data With Noisy Labels. [Paper]
  • Robust Long-Tailed Learning Under Label Noise. [Paper]
  • Instance-dependent Label-noise Learning under a Structural Causal Model. [Paper]
  • Assessing the Quality of the Datasets by Identifying Mislabeled Samples. [Paper]
  • Learning to Aggregate and Refine Noisy Labels for Visual Sentiment Analysis. [Paper]
  • Assessing the Quality of the Datasets by Identifying Mislabeled Samples. [Paper]

Papers & Code in 2020


ICML 2020

  • [UCSC REAL Lab] Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates. [Paper][Code 1] [Code 2]
  • Normalized Loss Functions for Deep Learning with Noisy Labels. [Paper][Code]
  • SIGUA: Forgetting May Make Learning with Noisy Labels More Robust. [Paper][Code]
  • Error-Bounded Correction of Noisy Labels. [Paper][Code]
  • Training Binary Neural Networks through Learning with Noisy Supervision. [Paper][Code]
  • Improving generalization by controlling label-noise information in neural network weights. [Paper][Code]
  • Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training. [Paper][Code]
  • Searching to Exploit Memorization Effect in Learning with Noisy Labels. [Paper][Code]
  • Learning with Bounded Instance and Label-dependent Label Noise. [Paper]
  • Label-Noise Robust Domain Adaptation. [Paper]
  • Beyond Synthetic Noise: Deep Learning on Controlled Noisy Labels. [Paper]
  • Does label smoothing mitigate label noise?. [Paper]
  • Learning with Multiple Complementary Labels. [Paper]
  • Deep k-NN for Noisy Labels. [Paper]
  • Extreme Multi-label Classification from Aggregated Labels. [Paper]

ICLR 2020

  • DivideMix: Learning with Noisy Labels as Semi-supervised Learning. [Paper][Code]
  • Learning from Rules Generalizing Labeled Exemplars. [Paper] [Code]
  • Robust training with ensemble consensus. [Paper][Code]
  • Self-labelling via simultaneous clustering and representation learning. [Paper][Code]
  • Can gradient clipping mitigate label noise? [Paper][Code]
  • Mutual Mean-Teaching: Pseudo Label Refinery for Unsupervised Domain Adaptation on Person Re-identification. [Paper][Code]
  • Curriculum Loss: Robust Learning and Generalization against Label Corruption. [Paper]
  • Simple and Effective Regularization Methods for Training on Noisily Labeled Data with Generalization Guarantee. [Paper]
  • SELF: Learning to Filter Noisy Labels with Self-Ensembling. [Paper]

Nips 2020

  • Part-dependent Label Noise: Towards Instance-dependent Label Noise. [Paper][Code]
  • Identifying Mislabeled Data using the Area Under the Margin Ranking. [Paper][Code]
  • Dual T: Reducing Estimation Error for Transition Matrix in Label-noise Learning. [Paper]
  • Early-Learning Regularization Prevents Memorization of Noisy Labels. [Paper][Code]
  • Coresets for Robust Training of Deep Neural Networks against Noisy Labels. [Paper][Code]
  • Modeling Noisy Annotations for Crowd Counting. [Paper][Code]
  • Robust Optimization for Fairness with Noisy Protected Groups. [Paper][Code]
  • Stochastic Optimization with Heavy-Tailed Noise via Accelerated Gradient Clipping. [Paper][Code]
  • A Topological Filter for Learning with Label Noise. [Paper][Code]
  • Self-Adaptive Training: beyond Empirical Risk Minimization. [Paper][Code]
  • Disentangling Human Error from the Ground Truth in Segmentation of Medical Images. [Paper][Code]
  • Non-Convex SGD Learns Halfspaces with Adversarial Label Noise. [Paper]
  • Efficient active learning of sparse halfspaces with arbitrary bounded noise. [Paper]
  • Semi-Supervised Partial Label Learning via Confidence-Rated Margin Maximization. [Paper]
  • Labelling unlabelled videos from scratch with multi-modal self-supervision. [Paper][Code]
  • Distribution Aligning Refinery of Pseudo-label for Imbalanced Semi-supervised Learning. [Paper][Code]
  • MetaPoison: Practical General-purpose Clean-label Data Poisoning. [Paper][Code 1][Code 2]
  • Provably Consistent Partial-Label Learning. [Paper]
  • A Variational Approach for Learning from Positive and Unlabeled Data. [Paper][Code]

AAAI 2020

  • [UCSC REAL Lab] Reinforcement Learning with Perturbed Rewards. [Paper] [Code]
  • Less Is Better: Unweighted Data Subsampling via Influence Function. [Paper] [Code]
  • Weakly Supervised Sequence Tagging from Noisy Rules. [Paper][Code]
  • Coupled-View Deep Classifier Learning from Multiple Noisy Annotators. [Paper]
  • Partial multi-label learning with noisy label identification. [Paper]
  • Self-Paced Robust Learning for Leveraging Clean Labels in Noisy Data. [Paper]
  • Label Error Correction and Generation Through Label Relationships. [Paper]

CVPR 2020

  • Combating noisy labels by agreement: A joint training method with co-regularization. [Paper][Code]
  • Distilling Effective Supervision From Severe Label Noise. [Paper][Code]
  • Self-Training With Noisy Student Improves ImageNet Classification. [Paper][Code]
  • Noise Robust Generative Adversarial Networks. [Paper][Code]
  • Global-Local GCN: Large-Scale Label Noise Cleansing for Face Recognition. [Paper]
  • DLWL: Improving Detection for Lowshot Classes With Weakly Labelled Data. [Paper]
  • Spherical Space Domain Adaptation With Robust Pseudo-Label Loss. [Paper][Code]
  • Training Noise-Robust Deep Neural Networks via Meta-Learning. [Paper][Code]
  • Shoestring: Graph-Based Semi-Supervised Classification With Severely Limited Labeled Data. [Paper][Code]
  • Noise-Aware Fully Webly Supervised Object Detection. [Paper][Code]
  • Learning From Noisy Anchors for One-Stage Object Detection. [Paper][Code]
  • Generating Accurate Pseudo-Labels in Semi-Supervised Learning and Avoiding Overconfident Predictions via Hermite Polynomial Activations. [Paper][Code]
  • Revisiting Knowledge Distillation via Label Smoothing Regularization. [Paper][Code]

ECCV 2020

  • 2020-ECCV - Learning with Noisy Class Labels for Instance Segmentation. [Paper][Code]
  • 2020-ECCV - Suppressing Mislabeled Data via Grouping and Self-Attention. [Paper][Code]
  • 2020-ECCV - NoiseRank: Unsupervised Label Noise Reduction with Dependence Models. [Paper]
  • 2020-ECCV - Weakly Supervised Learning with Side Information for Noisy Labeled Images. [Paper]
  • 2020-ECCV - Learning Noise-Aware Encoder-Decoder from Noisy Labels by Alternating Back-Propagation for Saliency Detection. [Paper]
  • 2020-ECCV - Graph convolutional networks for learning with few clean and many noisy labels. [Paper]

ArXiv 2020

  • No Regret Sample Selection with Noisy Labels. [Paper][Code]
  • Meta Soft Label Generation for Noisy Labels. [Paper][Code]
  • Learning from Noisy Labels with Deep Neural Networks: A Survey. [Paper]
  • RAR-U-Net: a Residual Encoder to Attention Decoder by Residual Connections Framework for Spine Segmentation under Noisy Labels. [Paper]
  • Learning from Small Amount of Medical Data with Noisy Labels: A Meta-Learning Approach. [Paper]

Owner
Jiaheng Wei
Ph.D@ UCSC CSE
Jiaheng Wei
PASTRIE: A Corpus of Prepositions Annotated with Supersense Tags in Reddit International English

PASTRIE Official release of the corpus described in the paper: Michael Kranzlein, Emma Manning, Siyao Peng, Shira Wein, Aryaman Arora, and Nathan Schn

NERT @ Georgetown 4 Dec 02, 2021
Neural Surface Maps

Neural Surface Maps Official implementation of Neural Surface Maps - Luca Morreale, Noam Aigerman, Vladimir Kim, Niloy J. Mitra [Paper] [Project Page]

Luca Morreale 49 Dec 13, 2022
A denoising autoencoder + adversarial losses and attention mechanisms for face swapping.

faceswap-GAN Adding Adversarial loss and perceptual loss (VGGface) to deepfakes'(reddit user) auto-encoder architecture. Updates Date Update 2018-08-2

3.2k Dec 30, 2022
Implementation of Axial attention - attending to multi-dimensional data efficiently

Axial Attention Implementation of Axial attention in Pytorch. A simple but powerful technique to attend to multi-dimensional data efficiently. It has

Phil Wang 250 Dec 25, 2022
Tensor-based approaches for fMRI classification

tensor-fmri Using tensor-based approaches to classify fMRI data from StarPLUS. Citation If you use any code in this repository, please cite the follow

4 Sep 07, 2022
AdelaiDepth is an open source toolbox for monocular depth prediction.

AdelaiDepth is an open source toolbox for monocular depth prediction.

Adelaide Intelligent Machines (AIM) Group 743 Jan 01, 2023
HAR-stacked-residual-bidir-LSTMs - Deep stacked residual bidirectional LSTMs for HAR

HAR-stacked-residual-bidir-LSTM The project is based on this repository which is presented as a tutorial. It consists of Human Activity Recognition (H

Guillaume Chevalier 287 Dec 27, 2022
Code for the paper "Regularizing Variational Autoencoder with Diversity and Uncertainty Awareness"

DU-VAE This is the pytorch implementation of the paper "Regularizing Variational Autoencoder with Diversity and Uncertainty Awareness" Acknowledgement

Dazhong Shen 4 Oct 19, 2022
clustimage is a python package for unsupervised clustering of images.

clustimage The aim of clustimage is to detect natural groups or clusters of images. Image recognition is a computer vision task for identifying and ve

Erdogan Taskesen 52 Jan 02, 2023
GB-CosFace: Rethinking Softmax-based Face Recognition from the Perspective of Open Set Classification

GB-CosFace: Rethinking Softmax-based Face Recognition from the Perspective of Open Set Classification This is the official pytorch implementation of t

Alibaba Cloud 5 Nov 14, 2022
The code is for the paper "A Self-Distillation Embedded Supervised Affinity Attention Model for Few-Shot Segmentation"

SD-AANet The code is for the paper "A Self-Distillation Embedded Supervised Affinity Attention Model for Few-Shot Segmentation" [arxiv] Overview confi

cv516Buaa 9 Nov 07, 2022
Source Code For Template-Based Named Entity Recognition Using BART

Template-Based NER Source Code For Template-Based Named Entity Recognition Using BART Training Training train.py Inference inference.py Corpus ATIS (h

174 Dec 19, 2022
PyTorch Implementation of Meta-StyleSpeech : Multi-Speaker Adaptive Text-to-Speech Generation

StyleSpeech - PyTorch Implementation PyTorch Implementation of Meta-StyleSpeech : Multi-Speaker Adaptive Text-to-Speech Generation. Status (2021.06.13

Keon Lee 140 Dec 21, 2022
True Few-Shot Learning with Language Models

This codebase supports using language models (LMs) for true few-shot learning: learning to perform a task using a limited number of examples from a single task distribution.

Ethan Perez 124 Jan 04, 2023
Safe Bayesian Optimization

SafeOpt - Safe Bayesian Optimization This code implements an adapted version of the safe, Bayesian optimization algorithm, SafeOpt [1], [2]. It also p

Felix Berkenkamp 111 Dec 11, 2022
a Pytorch easy re-implement of "YOLOX: Exceeding YOLO Series in 2021"

A pytorch easy re-implement of "YOLOX: Exceeding YOLO Series in 2021" 1. Notes This is a pytorch easy re-implement of "YOLOX: Exceeding YOLO Series in

91 Dec 26, 2022
Inflated i3d network with inception backbone, weights transfered from tensorflow

I3D models transfered from Tensorflow to PyTorch This repo contains several scripts that allow to transfer the weights from the tensorflow implementat

Yana 479 Dec 08, 2022
Modifications of the official PyTorch implementation of StyleGAN3. Let's easily generate images and videos with StyleGAN2/2-ADA/3!

Alias-Free Generative Adversarial Networks (StyleGAN3) Official PyTorch implementation of the NeurIPS 2021 paper Alias-Free Generative Adversarial Net

Diego Porres 185 Dec 24, 2022
Basics of 2D and 3D Human Pose Estimation.

Human Pose Estimation 101 If you want a slightly more rigorous tutorial and understand the basics of Human Pose Estimation and how the field has evolv

Sudharshan Chandra Babu 293 Dec 14, 2022
An example of time series augmentation methods with Keras

Time Series Augmentation This is a collection of time series data augmentation methods and an example use using Keras. News 2020/04/16: Repository Cre

九州大学 ヒューマンインタフェース研究室 229 Jan 02, 2023