All Projects → bnu-wangxun → Deep_metric

bnu-wangxun / Deep_metric

Licence: apache-2.0
Deep Metric Learning

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Deep metric

Hardnet
Hardnet descriptor model - "Working hard to know your neighbor's margins: Local descriptor learning loss"
Stars: ✭ 350 (-47.84%)
Mutual labels:  image-retrieval
Mimicry
[CVPR 2020 Workshop] A PyTorch GAN library that reproduces research results for popular GANs.
Stars: ✭ 458 (-31.74%)
Mutual labels:  cvpr
Hashing Baseline For Image Retrieval
Various hashing methods for image retrieval and serves as the baselines
Stars: ✭ 511 (-23.85%)
Mutual labels:  image-retrieval
Awesome deep learning interpretability
深度学习近年来关于神经网络模型解释性的相关高引用/顶会论文(附带代码)
Stars: ✭ 401 (-40.24%)
Mutual labels:  cvpr
Sis
Simple image search engine
Stars: ✭ 438 (-34.72%)
Mutual labels:  image-retrieval
Epipolarpose
Self-Supervised Learning of 3D Human Pose using Multi-view Geometry (CVPR2019)
Stars: ✭ 477 (-28.91%)
Mutual labels:  cvpr
Mvpose
Code for "Fast and Robust Multi-Person 3D Pose Estimation from Multiple Views" in CVPR'19
Stars: ✭ 326 (-51.42%)
Mutual labels:  cvpr
Zooming Slow Mo Cvpr 2020
Fast and Accurate One-Stage Space-Time Video Super-Resolution (accepted in CVPR 2020)
Stars: ✭ 555 (-17.29%)
Mutual labels:  cvpr
Cvpr2021 Papers With Code
CVPR 2021 论文和开源项目合集
Stars: ✭ 7,138 (+963.79%)
Mutual labels:  cvpr
Demon
DeMoN: Depth and Motion Network
Stars: ✭ 501 (-25.34%)
Mutual labels:  cvpr
Cbir
🏞 A content-based image retrieval (CBIR) system
Stars: ✭ 407 (-39.34%)
Mutual labels:  image-retrieval
Class Balanced Loss
Class-Balanced Loss Based on Effective Number of Samples. CVPR 2019
Stars: ✭ 433 (-35.47%)
Mutual labels:  cvpr
Deepalignmentnetwork
A deep neural network for face alignment
Stars: ✭ 480 (-28.46%)
Mutual labels:  cvpr
Fbrs interactive segmentation
[CVPR2020] f-BRS: Rethinking Backpropagating Refinement for Interactive Segmentation https://arxiv.org/abs/2001.10331
Stars: ✭ 366 (-45.45%)
Mutual labels:  cvpr
Cvpr 2019 Paper Statistics
Statistics and Visualization of acceptance rate, main keyword of CVPR 2019 accepted papers for the main Computer Vision conference (CVPR)
Stars: ✭ 527 (-21.46%)
Mutual labels:  cvpr
Mmt
[ICLR-2020] Mutual Mean-Teaching: Pseudo Label Refinery for Unsupervised Domain Adaptation on Person Re-identification.
Stars: ✭ 345 (-48.58%)
Mutual labels:  image-retrieval
Fsa Net
[CVPR19] FSA-Net: Learning Fine-Grained Structure Aggregation for Head Pose Estimation from a Single Image
Stars: ✭ 469 (-30.1%)
Mutual labels:  cvpr
Cnn For Image Retrieval
🌅The code of post "Image retrieval using MatconvNet and pre-trained imageNet"
Stars: ✭ 597 (-11.03%)
Mutual labels:  image-retrieval
Cv paperdaily
CV 论文笔记
Stars: ✭ 555 (-17.29%)
Mutual labels:  cvpr
Awesome Cvpr Paper
CVPR 论文收集,包含但不限于2021、2020、2019、2018、2017文章
Stars: ✭ 493 (-26.53%)
Mutual labels:  cvpr

Deep Metric Learning in PyTorch

 Learn deep metric for image retrieval or other information retrieval. 

Our XBM is nominated as best paper in CVPR 2020.

One Blog on XBM in Zhihu

我写了一个知乎文章,通俗快速解读了XBM想法动机:

跨越时空的难样本挖掘

欢迎大家阅读指点!

Recommend one recently released excellent papers in DML not written by me:

A Metric Learning Reality Check

from Cornell Tech and Facebook AI

Abstract: Deep metric learning papers from the past four years have consistently claimed great advances in accuracy, often more than doubling the performance of decade-old methods. In this paper, we take a closer look at the field to see if this is actually true. We find flaws in the experimental setup of these papers, and propose a new way to evaluate metric learning algorithms. Finally, we present experimental results that show that the improvements over time have been marginal at best.

  

XBM: A New Sota method for DML, accepted by CVPR-2020 as Oral and nominated as best paper:

Cross-Batch Memory for Embedding Learning

  • Great Improvement: XBM can improve the [email protected] by 12~25% on three large-scale datasets

  • Easy to implement: with only several lines of codes

  • Memory efficient: with less than 1GB for large-scale datasets

  • Code has already been released: xbm

Other implementations:

pytorch-metric-learning(a great work by Kevin Musgrave)

     

MS Loss based on GPW: Accepted by CVPR 2019 as Poster

Multi-Similarity Loss with General Pair Weighting for Deep Metric Learning

 

Deep metric methods implemented in this repositories:

  • Contrasstive Loss [1]

  • Semi-Hard Mining Strategy [2]

  • Lifted Structure Loss* [3] (Modified version because of its original weak performance)

  • Binomial BinDeviance Loss [4]

  • NCA Loss [6]

  • Multi-Similarity Loss [7]

Dataset

  • Car-196

    first 98 classes as train set and last 98 classes as test set

  • CUB-200-2011

    first 100 classes as train set and last 100 classes as test set

  • Stanford-Online-Products

    for the experiments, we split 59,551 images of 11,318 classes for training and 60,502 images of 11,316 classes for testing

  • In-Shop-clothes-Retrieval

    For the In-Shop Clothes Retrieval dataset, 3,997 classes with 25,882 images for training. And the test set are partitioned to query set with 3,985 classes(14,218 images) and gallery set with 3,985 classes (12,612 images).

  • Processed CUB and Cars196

    Extract code: inmj

    To easily reimplement the performance, I provide the processed datasets: CUB and Cars-196.

Requirements

  • Python >= 3.5
  • PyTorch = 1.0

Comparasion with state-of-the-art on CUB-200 and Cars-196

[email protected] 1 2 4 8 16 32 1 2 4 8 16 32
HDC 53.6 65.7 77.0 85.6 91.5 95.5 73.7 83.2 89.5 93.8 96.7 98.4
Clustering 48.2 61.4 71.8 81.9 - - 58.1 70.6 80.3 87.8 - -
ProxyNCA 49.2 61.9 67.9 72.4 - - 73.2 82.4 86.4 87.8 - -
Smart Mining 49.8 62.3 74.1 83.3 - - 64.7 76.2 84.2 90.2 - -
Margin [5] 63.6 74.4 83.1 90.0 94.2 - 79.6 86.5 91.9 95.1 97.3 -
HTL 57.1 68.8 78.7 86.5 92.5 95.5 81.4 88.0 92.7 95.7 97.4 99.0
ABIER 57.5 68.7 78.3 86.2 91.9 95.5 82.0 89.0 93.2 96.1 97.8 98.7

Comparasion with state-of-the-art on SOP and In-shop

[email protected] 1 10 100 1000 1 10 20 30 40 50
Clustering 67.0 83.7 93.2 - - - - - - -
HDC 69.5 84.4 92.8 97.7 62.1 84.9 89.0 91.2 92.3 93.1
Margin [5] 72.7 86.2 93.8 98.0 - - - - - -
Proxy-NCA 73.7 - - - - - - - - -
ABIER 74.2 86.9 94.0 97.8 83.1 95.1 96.9 97.5 97.8 98.0
HTL 74.8 88.3 94.8 98.4 80.9 94.3 95.8 97.2 97.4 97.8

see more detail in our CVPR-2019 paper Multi-Similarity Loss

Reproducing Car-196 (or CUB-200-2011) experiments

*** weight :***

sh run_train_00.sh

Other implementations:

[Tensorflow] (by geonm)

References

[1] [R. Hadsell, S. Chopra, and Y. LeCun. Dimensionality reduction by learning an invariant mapping]

[2] [F. Schroff, D. Kalenichenko, and J. Philbin. Facenet: A unified embedding for face recognition and clustering. In CVPR, 2015.]

[3][H. Oh Song, Y. Xiang, S. Jegelka, and S. Savarese. Deep metric learning via lifted structured feature embedding. In CVPR, 2016.]

[4][D. Yi, Z. Lei, and S. Z. Li. Deep metric learning for practical person re-identification.]

[5][C. Wu, R. Manmatha, A. J. Smola, and P. Kr¨ahenb¨uhl. Sampling matters in deep embedding learning. ICCV, 2017.]

[6][R. Salakhutdinov and G. Hinton. Learning a nonlinear embedding by preserving class neighbourhood structure. In AISTATS, 2007.]

Citation

If you use this method or this code in your research, please cite as:

@inproceedings{wang2019multi,
title={Multi-Similarity Loss with General Pair Weighting for Deep Metric Learning},
author={Wang, Xun and Han, Xintong and Huang, Weilin and Dong, Dengke and Scott, Matthew R},
booktitle={CVPR},
year={2019}
}

@inproceedings{wang2020xbm,
title={Cross-Batch Memory for Embedding Learning},
author={Wang, Xun and Zhang, haozhi and Huang, Weilin and Scott, Matthew R},
booktitle={CVPR},
year={2020}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].