All Projects → cleardusk → EmotionChallenge

cleardusk / EmotionChallenge

Licence: other
Source code for 1st winner of face micro-emotion competition, FG 2017.

Programming Languages

python
139335 projects - #7 most used programming language
C++
36643 projects - #6 most used programming language
TeX
3793 projects

Projects that are alternatives of or similar to EmotionChallenge

converse
Conversational text Analysis using various NLP techniques
Stars: ✭ 147 (+297.3%)
Mutual labels:  emotion-recognition
adversarial-attacks
Code for our CVPR 2018 paper, "On the Robustness of Semantic Segmentation Models to Adversarial Attacks"
Stars: ✭ 90 (+143.24%)
Mutual labels:  caffe
CLUEmotionAnalysis2020
CLUE Emotion Analysis Dataset 细粒度情感分析数据集
Stars: ✭ 3 (-91.89%)
Mutual labels:  emotion-recognition
yolov5 onnx2caffe
yolov5 onnx caffe
Stars: ✭ 73 (+97.3%)
Mutual labels:  caffe
superpixelRefinement
Superpixel-based Refinement for Object Proposal Generation (ICPR 2020)
Stars: ✭ 24 (-35.14%)
Mutual labels:  caffe
Self-Supervised-Embedding-Fusion-Transformer
The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.
Stars: ✭ 57 (+54.05%)
Mutual labels:  emotion-recognition
deep-learning-platforms
deep-learning platforms,framework,data(深度学习平台、框架、资料)
Stars: ✭ 17 (-54.05%)
Mutual labels:  caffe
m3f.pytorch
PyTorch code for "M³T: Multi-Modal Multi-Task Learning for Continuous Valence-Arousal Estimation"
Stars: ✭ 20 (-45.95%)
Mutual labels:  emotion-recognition
hfusion
Multimodal sentiment analysis using hierarchical fusion with context modeling
Stars: ✭ 42 (+13.51%)
Mutual labels:  emotion-recognition
loki
Proof-of-concept of emotion-targeted content delivery using machine learning and ARKit.
Stars: ✭ 76 (+105.41%)
Mutual labels:  emotion-recognition
TrainCaffeCustomDataset
Transfer learning in Caffe: example on how to train CaffeNet on custom dataset
Stars: ✭ 20 (-45.95%)
Mutual labels:  caffe
ppq
PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.
Stars: ✭ 281 (+659.46%)
Mutual labels:  caffe
AIML-Human-Attributes-Detection-with-Facial-Feature-Extraction
This is a Human Attributes Detection program with facial features extraction. It detects facial coordinates using FaceNet model and uses MXNet facial attribute extraction model for extracting 40 types of facial attributes. This solution also detects Emotion, Age and Gender along with facial attributes.
Stars: ✭ 48 (+29.73%)
Mutual labels:  emotion-recognition
Deeplearning-digital-pathology
Full package for applying deep learning to virtual slides.
Stars: ✭ 59 (+59.46%)
Mutual labels:  caffe
Caffe-installation-Raspberry-Pi-3
No description or website provided.
Stars: ✭ 23 (-37.84%)
Mutual labels:  caffe
STEP
Spatial Temporal Graph Convolutional Networks for Emotion Perception from Gaits
Stars: ✭ 39 (+5.41%)
Mutual labels:  emotion-recognition
soxan
Wav2Vec for speech recognition, classification, and audio classification
Stars: ✭ 113 (+205.41%)
Mutual labels:  emotion-recognition
Age-gender-and-emotion-recognition
3 networks to recognition age,gender and emotion
Stars: ✭ 29 (-21.62%)
Mutual labels:  emotion-recognition
ntua-slp-semeval2018
Deep-learning models of NTUA-SLP team submitted in SemEval 2018 tasks 1, 2 and 3.
Stars: ✭ 79 (+113.51%)
Mutual labels:  emotion-recognition
theWorldInSafety
Surveillance System Against Violence
Stars: ✭ 31 (-16.22%)
Mutual labels:  caffe

Citation

If you use these models or code in your research, please cite:

@inproceedings{guo2017multi,
  title={Multi-modality Network with Visual and Geometrical Information for Micro Emotion Recognition},
  author={Guo, Jianzhu and Zhou, Shuai and Wu, Jinlin and Wan, Jun and Zhu, Xiangyu and Lei, Zhen and Li, Stan Z},
  booktitle={Automatic Face \& Gesture Recognition (FG 2017), 2017 12th IEEE International Conference on},
  pages={814--819},
  year={2017},
  organization={IEEE}
}

@article{guo2018dominant,
  title={Dominant and Complementary Emotion Recognition from Still Images of Faces},
  author={Guo, Jianzhu and Lei, Zhen and Wan, Jun and Avots, Egils and Hajarolasvadi, Noushin and Knyazev, Boris and Kuharenko, Artem and Jacques, Julio CS and Bar{\'o}, Xavier and Demirel, Hasan and others},
  journal={IEEE Access},
  year={2018},
  publisher={IEEE}
}

For final evaluation

Submission

First, you should generate the crop and aligned data on test Chanllenge dataset. Just change to crop_align dir

python landmark.py

The crop and aligned of test data(final evaluation phase data) of 224x224 will place in $ROOT/data/face_224

Then change to cnn dir, just type

python extract.py

It will load data preprocessed and caffe model to generate labels named predictions.txt and predictions.zip for test data. All details were considered.

In this repo, some directory path may be confused, just be careful, contact me if any questions occured.

The trained caffe model is just a experiement model, it may not has the best perfomance in this challenge.

Just upload predictions.zip to submit window then.

Introduction

We use Dlib to do face and landmark detection, and use landmark to do face cropping and alignment, then we use Caffe to with landmark and cropping image to train a cnn model to do the face expression recognition task.

Pipline

Preproces

First, run landmark.py to get all the origin image landmark, then build the crop_align binary, and run crop_align.py to get all the 224x224 size image.

Build crop_align

cd crop_align
mkdir build
cd build
cmake ..
make

All the preprocessed data except the images are in data dir.

Training

Change to cnn dir, run prepare_data.py to prepare training, validation and test data. Then run train_val.sh to start training.

Extract(Test)

Just run extract.py to generate the result, the input is the test image and its landmark offset info.

Method

We use the landmark offset and image info to do this task. In detail, the landmark offset is calculated by substraction of 224x224 image landmark and each id's mean landmark, and we concact this feature to modified alexnet's last output feature. We change softmax loss to hinge loss to get a little better result.

More detail is in the fact_sheet.tex.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].