All Projects → WeitaoVan → L Gm Loss

WeitaoVan / L Gm Loss

Licence: mit
Implementation of our accepted CVPR 2018 paper "Rethinking Feature Distribution for Loss Functions in Image Classification"

Labels

Projects that are alternatives of or similar to L Gm Loss

Awesome Cvpr Paper
CVPR 论文收集,包含但不限于2021、2020、2019、2018、2017文章
Stars: ✭ 493 (+204.32%)
Mutual labels:  cvpr2018
Vc
Code for CVPR'18 "Grounding Referring Expressions in Images by Variational Context"
Stars: ✭ 26 (-83.95%)
Mutual labels:  cvpr2018
Dhf1k
Revisiting Video Saliency: A Large-scale Benchmark and a New Model (CVPR18, PAMI19)
Stars: ✭ 96 (-40.74%)
Mutual labels:  cvpr2018
Lemniscate.pytorch
Unsupervised Feature Learning via Non-parametric Instance Discrimination
Stars: ✭ 532 (+228.4%)
Mutual labels:  cvpr2018
Polyrnn Pp
Inference Code for Polygon-RNN++ (CVPR 2018)
Stars: ✭ 704 (+334.57%)
Mutual labels:  cvpr2018
Facekit
Implementations of PCN (an accurate real-time rotation-invariant face detector) and other face-related algorithms
Stars: ✭ 1,028 (+534.57%)
Mutual labels:  cvpr2018
Atlasnet
This repository contains the source codes for the paper "AtlasNet: A Papier-Mâché Approach to Learning 3D Surface Generation ". The network is able to synthesize a mesh (point cloud + connectivity) from a low-resolution point cloud, or from an image.
Stars: ✭ 450 (+177.78%)
Mutual labels:  cvpr2018
Dcnets
Implementation for <Decoupled Networks> in CVPR'18.
Stars: ✭ 115 (-29.01%)
Mutual labels:  cvpr2018
Ffnet
Implementation of our CVPR 2018 Paper (FFNet: Video Fast-Forwarding via Reinforcement Learning)
Stars: ✭ 17 (-89.51%)
Mutual labels:  cvpr2018
Surfacenetworks
Source code for CVPR 2018 Oral paper "Surface Networks"
Stars: ✭ 83 (-48.77%)
Mutual labels:  cvpr2018
Stargan
StarGAN - Official PyTorch Implementation (CVPR 2018)
Stars: ✭ 4,946 (+2953.09%)
Mutual labels:  cvpr2018
Polyrnn Pp Pytorch
PyTorch training/tool code for Polygon-RNN++ (CVPR 2018)
Stars: ✭ 672 (+314.81%)
Mutual labels:  cvpr2018
Pwc Net
PWC-Net: CNNs for Optical Flow Using Pyramid, Warping, and Cost Volume, CVPR 2018 (Oral)
Stars: ✭ 1,142 (+604.94%)
Mutual labels:  cvpr2018
Fots.pytorch
FOTS Pytorch Implementation
Stars: ✭ 513 (+216.67%)
Mutual labels:  cvpr2018
Idn Caffe
Caffe implementation of "Fast and Accurate Single Image Super-Resolution via Information Distillation Network" (CVPR 2018)
Stars: ✭ 104 (-35.8%)
Mutual labels:  cvpr2018
Liteflownet
LiteFlowNet: A Lightweight Convolutional Neural Network for Optical Flow Estimation, CVPR 2018 (Spotlight paper, 6.6%)
Stars: ✭ 474 (+192.59%)
Mutual labels:  cvpr2018
Lab
[CVPR 2018] Look at Boundary: A Boundary-Aware Face Alignment Algorithm
Stars: ✭ 956 (+490.12%)
Mutual labels:  cvpr2018
Frvsr
Frame-Recurrent Video Super-Resolution (official repository)
Stars: ✭ 157 (-3.09%)
Mutual labels:  cvpr2018
Residualdensenetwork Pytorch
Pytorch implement: Residual Dense Network for Image Super-Resolution
Stars: ✭ 107 (-33.95%)
Mutual labels:  cvpr2018
Super Slowmo
An attempt at a PyTorch implimentation of "Super SloMo: High Quality Estimation of Multiple Intermediate Frames for Video Interpolation"
Stars: ✭ 73 (-54.94%)
Mutual labels:  cvpr2018

L-GM-loss

For Caffe and Tensorflow.
Implementation of our CVPR 2018 paper "Rethinking Feature Distribution for Loss Functions in Image Classification".
Paper authors: Weitao Wan, Yuanyi Zhong, Tianpeng Li, Jiansheng Chen.

Experiments in our paper were carried out through the Caffe implementation.
The folder tensorflow contains the Tensorflow demo.

Code is written by Yuanyi Zhong and Weitao Wan.

Abstract

We propose a large-margin Gaussian Mixture (L-GM) loss for deep neural networks in classification tasks. Different from the softmax cross-entropy loss, our proposal is established on the assumption that the deep features of the training set follow a Gaussian Mixture distribution. By involving a classification margin and a likelihood regularization, the L-GM loss facilitates both a high classification performance and an accurate modeling of the training feature distribution. As such, the L-GM loss is superior to the softmax loss and its major variants in the sense that besides classification, it can be readily used to distinguish abnormal inputs, such as the adversarial examples, based on their features' likelihood to the training feature distribution. Extensive experiments on various recognition benchmarks like MNIST, CIFAR, ImageNet and LFW, as well as on adversarial examples demonstrate the effectiveness of our proposal.

Instructions

(For tensorflow, please enter the tensorflow folder)

  ./train.sh 0 simple  # 0 is the GPU id, simple is the folder containing network definitions and solver

Layer details

  • Specify margin parameter α and likelihood weight λ, which is margin_mul and center_coef in the layer param, respectively.
    margin_mul {
      policy: STEPUP
      value: 0.1
      step: 5000 
      gamma: 2
      max: 0.3 
}

This specifies a gradually growing value for α (multiplied by 2 every 5000 iterations, with initial value 0.1 and final maximum value 0.3), which is helpful for training.

  • other indicators
update_sigma: false

Fix the variances to initial values (1.0).

isotropic: true

The variances of different dimensions are identical.

Data

We've described how the data is pre-processed in our paper. For example, the CIFAR-100 training data (32x32) is padded to 40x40 with zero pixels and then randomly cropped with a 32x32 window for training.
In the CIFAR-100 example, we use data in HDF5 format. You can choose other formats, changing the data layer accordingly.

The CIFAR100 training data (with or without augmentation) and test data can be downloaded from Baidu Drive(.h5 file).

Citations

If you find this work useful, please consider citing it.

@inproceedings{LGM2018,
  title={Rethinking Feature Distribution for Loss Functions in Image Classification},
  author={Wan, Weitao and Zhong, Yuanyi and Li, Tianpeng and Chen, Jiansheng},
  booktitle={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2018}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].