All Projects → ColumbiaDVMM → Heated_Up_Softmax_Embedding

ColumbiaDVMM / Heated_Up_Softmax_Embedding

Licence: BSD-2-Clause License
Project page for Heated-up Softmax Embedding

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Heated Up Softmax Embedding

FastAP-metric-learning
Code for CVPR 2019 paper "Deep Metric Learning to Rank"
Stars: ✭ 93 (+121.43%)
Mutual labels:  metric-learning
LearningToCompare-Tensorflow
Tensorflow implementation for paper: Learning to Compare: Relation Network for Few-Shot Learning.
Stars: ✭ 17 (-59.52%)
Mutual labels:  metric-learning
ePillID-benchmark
ePillID Dataset: A Low-Shot Fine-Grained Benchmark for Pill Identification (CVPR 2020 VL3)
Stars: ✭ 54 (+28.57%)
Mutual labels:  metric-learning
GeDML
Generalized Deep Metric Learning.
Stars: ✭ 30 (-28.57%)
Mutual labels:  metric-learning
Deep-Learning-Mahjong---
Reinforcement learning (RL) implementation of imperfect information game Mahjong using markov decision processes to predict future game states
Stars: ✭ 45 (+7.14%)
Mutual labels:  softmax
NCE-loss
Tensorflow NCE loss in Keras
Stars: ✭ 30 (-28.57%)
Mutual labels:  softmax
S-WMD
Code for Supervised Word Mover's Distance (SWMD)
Stars: ✭ 90 (+114.29%)
Mutual labels:  metric-learning
Machine-Learning-in-Python-Workshop
My workshop on machine learning using python language to implement different algorithms
Stars: ✭ 89 (+111.9%)
Mutual labels:  softmax
Metric Learning Adversarial Robustness
Code for NeurIPS 2019 Paper
Stars: ✭ 44 (+4.76%)
Mutual labels:  metric-learning
MinkLoc3D
MinkLoc3D: Point Cloud Based Large-Scale Place Recognition
Stars: ✭ 83 (+97.62%)
Mutual labels:  metric-learning
Deep-Metric-Embedding
Papers and Codes about Deep Metric Learning/Deep Embedding
Stars: ✭ 37 (-11.9%)
Mutual labels:  metric-learning
lfda
Local Fisher Discriminant Analysis in R
Stars: ✭ 74 (+76.19%)
Mutual labels:  metric-learning
recipe1m.bootstrap.pytorch
Retrieve recipes from foodie pictures using Deep Learning and Pytorch
Stars: ✭ 42 (+0%)
Mutual labels:  metric-learning
TCE
This repository contains the code implementation used in the paper Temporally Coherent Embeddings for Self-Supervised Video Representation Learning (TCE).
Stars: ✭ 51 (+21.43%)
Mutual labels:  metric-learning
efficient softmax
BlackOut and Adaptive Softmax for language models by Chainer
Stars: ✭ 12 (-71.43%)
Mutual labels:  softmax
TreeRep
Learning Tree structures and Tree metrics
Stars: ✭ 18 (-57.14%)
Mutual labels:  metric-learning
scLearn
scLearn:Learning for single cell assignment
Stars: ✭ 26 (-38.1%)
Mutual labels:  metric-learning
dml
R package for Distance Metric Learning
Stars: ✭ 58 (+38.1%)
Mutual labels:  metric-learning
Learning-Lab-C-Library
This library provides a set of basic functions for different type of deep learning (and other) algorithms in C.This deep learning library will be constantly updated
Stars: ✭ 20 (-52.38%)
Mutual labels:  softmax
SPML
Universal Weakly Supervised Segmentation by Pixel-to-Segment Contrastive Learning
Stars: ✭ 81 (+92.86%)
Mutual labels:  metric-learning

Heated-Up Softmax Embedding

Project page for Heated-Up Softmax Embedding. We propose a “heating-up” strategy to train a classifier with increasing temperatures, leading the corresponding embeddings to achieve state-of-the-art performance on a variety of metric learning benchmarks.

Introduction

Descriptor extraction is mapping an image to a point in the descriptor space.

A good descriptor should be:

  • Compact: Samples from the same class to be close
  • Spread-out: Samples from different classes to be far away

Bottleneck feature from classification network shows strong performance in [1]

Bottleneck feature is not learned to be compact and spread-out, thus may not be suitable for clustering and retreival. Here are some features learned from MNIST. Each color shows one digit. The diamond shows classifier weight.

Temperature Parameter 𝛼

Applying both ℓ2 normalization to both feature and weight show strong performance in face verification [2,3]. Need temperature parameter 𝛼 for gradient tuning.

Here are some features learned from MNIST with new pipeline. 𝛼 plays an important role in determining the final distribution.

Consider the magnitude of the gradient with respect to descriptor, different 𝛼 values assign different gradients to different samples.

Heated-Up Feature

At the begining of the training, we need intermediate 𝛼 to assign large gradient to incorrect sample and medium gradient to boundary sample. If 𝛼 is too large (red dash line to the right hand side), boundary sample may not get enough gradient to update, the feature will be not compact. If 𝛼 is too small (red dash line to the left hand side, close to 0), all samples will have similar gradient, training may not be effective (no "hard mining").

At the end of the training, all the samples are correctly classified. We need small 𝛼 to assign large gradient to all the samples to further compress the distribution.

Requirements

See ./requirements.txt

Run the code

  1. Prepare the data. Run codes in ./dataset/ to download and preprocess the data.
python ./dataset/car196.py -h
  1. Download inception_v1 from Tensorflow site, and save it in ./data/model/.

  2. Check ./tensorflow/run_Inception.py to see how to run on Car196, Bird200 and Product datasets.

python ./tensorflow/run_Inception.py
  1. Check ./tensorflow/run_mnist_lenet.py to see how to run on Mnist dataset to get the distribution map.
python ./tensorflow/run_mnist_lenet.py

Citation

Please cite us if you find this code useful:

@inproceedings{zhang2018Heated,
title={Heated-Up Softmax Embedding},
author={Zhang, Xu and Yu, Felix X. and Karaman, Svebor and Zhang, Wei and Chang, Shih-Fu},
booktitle={arXiv preprint arXiv:1809.04157},
year={2018}
}

Reference

[1] Y. Movshovitz-Attias, A. Toshev, T. K. Leung, S. Ioffe, and S. Singh. No Fuss Distance Metric Learning Using Proxies. In ICCV 2017

[2] Wang, Feng, Xiang Xiang, Jian Cheng, and Alan L. Yuille. NormFace: L2 Hypersphere Embedding for Face Verification. In ACM MM 2017

[3] Liu, Weiyang, Yandong Wen, Zhiding Yu, et al. SphereFace: Deep Hypersphere Embedding for Face Recognition. In CVPR 2017

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].