A Light CNN for Deep Face Representation with Noisy Labels

Overview

A Light CNN for Deep Face Representation with Noisy Labels

Citation

If you use our models, please cite the following paper:

@article{wulight,
  title={A Light CNN for Deep Face Representation with Noisy Labels},
  author={Wu, Xiang and He, Ran and Sun, Zhenan and Tan, Tieniu}
  journal={arXiv preprint arXiv:1511.02683},
  year={2015}
}
@article{wu2015lightened,
  title={A Lightened CNN for Deep Face Representation},
  author={Wu, Xiang and He, Ran and Sun, Zhenan},
  journal={arXiv preprint arXiv:1511.02683},
  year={2015}
}
@article{wu2015learning,
  title={Learning Robust Deep Face Representation},
  author={Wu, Xiang},
  journal={arXiv preprint arXiv:1507.04844},
  year={2015}
}

Updates

  • Dec 16, 2016
  • Nov 08, 2016
    • The prototxt and model C based on caffe-rc3 is updated. The accuracy on LFW achieves 98.80% and the [email protected]=0 obtains 94.97%.
    • The performance of set 1 on MegaFace achieves 65.532% for rank-1 accuracy and 75.854% for [email protected]=10^-6.
  • Nov 26, 2015
    • The prototxt and model B is updated and the accuracy on LFW achieves 98.13% for a single net without training on LFW.
  • Aug 13, 2015
    • Evaluation of LFW for identification protocols is published.
  • Jun 11, 2015
    • The prototxt and model A is released. The accuracy on LFW achieves 97.77%.

Overview

The Deep Face Representation Experiment is based on Convolution Neural Network to learn a robust feature for face verification task. The popular deep learning framework caffe is used for training on face datasets such as CASIA-WebFace, VGG-Face and MS-Celeb-1M. And the feature extraction is realized by python code caffe_ftr.py.

Structure

  • Code
    • data pre-processing and evaluation code
  • Model
    • caffemodel.
      • The model A and B is trained on CASIA-WebFace by caffe-rc.
      • The model C is trained on MS-Celeb-1M by caffe-rc3.
  • Proto
    • Lightened CNN implementations by caffe
  • Results
    • LFW features

Description

Data Pre-processing

  1. Download face dataset such as CASIA-WebFace, VGG-Face and MS-Celeb-1M.
  2. All face images are converted to gray-scale images and normalized to 144x144 according to landmarks.
  3. According to the 5 facial points, we not only rotate two eye points horizontally but also set the distance between the midpoint of eyes and the midpoint of mouth(ec_mc_y), and the y axis of midpoint of eyes(ec_y) .
Dataset size ec_mc_y ec_y
Training set 144x144 48 48
Testing set 128x128 48 40

Training

  1. The model is trained by open source deep learning framework caffe.
  2. The network configuration is showed in "proto" file and the trained model is showed in "model" file.

Evaluation

  1. The model is evaluated on LFW which is a popular data set for face verification task.
  2. The extracted features and lfw testing pairs are located in "results" file.
  3. To evaluate the model, the matlab code or other ROC evaluation code can be used.
  4. The model is also evaluated on MegaFace. The dataset and evaluation code can be downloaded from http://megaface.cs.washington.edu/

Results

The single convolution net testing is evaluated on unsupervised setting only computing cosine similarity for lfw pairs.

Model 100% - EER [email protected]=1% [email protected]=0.1% [email protected]=0 Rank-1 [email protected]=1%
A 97.77% 94.80% 84.37% 43.17% 84.79% 63.09%
B 98.13% 96.73% 87.13% 64.33% 89.21% 69.46%
C 98.80% 98.60% 96.77% 94.97% 93.80% 84.40%

The details are published as a technical report on arXiv.

The released models are only allowed for non-commercial use.

Owner
Alfred Xiang Wu
魔炮厨 | 夏娜厨 | 久远厨 | 珂朵莉厨 | PSN: wkira_vivio
Alfred Xiang Wu
Official implementation for paper Render In-between: Motion Guided Video Synthesis for Action Interpolation

Render In-between: Motion Guided Video Synthesis for Action Interpolation [Paper] [Supp] [arXiv] [4min Video] This is the official Pytorch implementat

8 Oct 27, 2022
The fundamental package for scientific computing with Python.

NumPy is the fundamental package needed for scientific computing with Python. Website: https://www.numpy.org Documentation: https://numpy.org/doc Mail

NumPy 22.4k Jan 09, 2023
Unsupervised Image Generation with Infinite Generative Adversarial Networks

Unsupervised Image Generation with Infinite Generative Adversarial Networks Here is the implementation of MICGANs using DCGAN architecture on MNIST da

16 Dec 24, 2021
PyTorch Implementation of ByteDance's Cross-speaker Emotion Transfer Based on Speaker Condition Layer Normalization and Semi-Supervised Training in Text-To-Speech

Cross-Speaker-Emotion-Transfer - PyTorch Implementation PyTorch Implementation of ByteDance's Cross-speaker Emotion Transfer Based on Speaker Conditio

Keon Lee 114 Jan 08, 2023
A PyTorch version of You Only Look at One-level Feature object detector

PyTorch_YOLOF A PyTorch version of You Only Look at One-level Feature object detector. The input image must be resized to have their shorter side bein

Jianhua Yang 25 Dec 30, 2022
Using Hotel Data to predict High Value And Potential VIP Guests

Description Using hotel data and AI to predict high value guests and potential VIP guests. Hotel can leverage on prediction resutls to run more effect

HCG 12 Feb 14, 2022
Advancing Self-supervised Monocular Depth Learning with Sparse LiDAR

Official implementation for paper "Advancing Self-supervised Monocular Depth Learning with Sparse LiDAR"

Ziyue Feng 72 Dec 09, 2022
MetaShift: A Dataset of Datasets for Evaluating Contextual Distribution Shifts and Training Conflicts (ICLR 2022)

MetaShift: A Dataset of Datasets for Evaluating Distribution Shifts and Training Conflicts This repo provides the PyTorch source code of our paper: Me

88 Jan 04, 2023
FlowTorch is a PyTorch library for learning and sampling from complex probability distributions using a class of methods called Normalizing Flows

FlowTorch is a PyTorch library for learning and sampling from complex probability distributions using a class of methods called Normalizing Flows.

Meta Incubator 272 Jan 02, 2023
This repository includes the official project for the paper: TransMix: Attend to Mix for Vision Transformers.

TransMix: Attend to Mix for Vision Transformers This repository includes the official project for the paper: TransMix: Attend to Mix for Vision Transf

Jie-Neng Chen 130 Jan 01, 2023
Depression Asisstant GDSC Challenge Solution

Depression Asisstant can help you give solution. Please using Python version 3.9.5 for contribute.

Ananda Rauf 1 Jan 30, 2022
AgML is a comprehensive library for agricultural machine learning

AgML is a comprehensive library for agricultural machine learning. Currently, AgML provides access to a wealth of public agricultural datasets for common agricultural deep learning tasks.

Plant AI and Biophysics Lab 1 Jul 07, 2022
deep learning model with only python and numpy with test accuracy 99 % on mnist dataset and different optimization choices

deep_nn_model_with_only_python_100%_test_accuracy deep learning model with only python and numpy with test accuracy 99 % on mnist dataset and differen

0 Aug 28, 2022
A PyTorch Implementation of "Neural Arithmetic Logic Units"

Neural Arithmetic Logic Units [WIP] This is a PyTorch implementation of Neural Arithmetic Logic Units by Andrew Trask, Felix Hill, Scott Reed, Jack Ra

Kevin Zakka 181 Nov 18, 2022
Fuzzing JavaScript Engines with Aspect-preserving Mutation

DIE Repository for "Fuzzing JavaScript Engines with Aspect-preserving Mutation" (in S&P'20). You can check the paper for technical details. Environmen

gts3.org (<a href=[email protected])"> 190 Dec 11, 2022
PyTorch implementation of UPFlow (unsupervised optical flow learning)

UPFlow: Upsampling Pyramid for Unsupervised Optical Flow Learning By Kunming Luo, Chuan Wang, Shuaicheng Liu, Haoqiang Fan, Jue Wang, Jian Sun Megvii

kunming luo 87 Dec 20, 2022
Pytorch-diffusion - A basic PyTorch implementation of 'Denoising Diffusion Probabilistic Models'

PyTorch implementation of 'Denoising Diffusion Probabilistic Models' This reposi

Arthur Juliani 76 Jan 07, 2023
Complete the code of prefix-tuning in low data setting

Prefix Tuning Note: 作者在论文中提到使用真实的word去初始化prefix的操作(Initializing the prefix with activations of real words,significantly improves generation)。我在使用作者提供的

Andrew Zeng 4 Jul 11, 2022
OneShot Learning-based hotword detection.

EfficientWord-Net Hotword detection based on one-shot learning Home assistants require special phrases called hotwords to get activated (eg:"ok google

ANT-BRaiN 102 Dec 25, 2022
Deep Q Learning with OpenAI Gym and Pokemon Showdown

pokemon-deep-learning An openAI gym project for pokemon involving deep q learning. Made by myself, Sam Little, and Layton Webber. This code captures g

2 Dec 22, 2021