All Projects → ZhengPeng7 → Sanet Keras

ZhengPeng7 / Sanet Keras

Licence: mit
Implement SANet for crowd counting in Keras.

Projects that are alternatives of or similar to Sanet Keras

Pydata Amsterdam 2016
Machine Learning with Scikit-Learn (material for pydata Amsterdam 2016)
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
Attentive Neural Processes
implementing "recurrent attentive neural processes" to forecast power usage (w. LSTM baseline, MCDropout)
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Multi Label
Pytorch code for multi-Instance multi-label problem
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Omx
Open Matrix (OMX)
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
Yolact Tutorial
A tutorial for using YOLACT in Google Colab
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Numerical methods youtube
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
2016learnpython
Python Teaching, Seminars for 2nd year students of School of Linguistics NRU HSE
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
Natural Language Processing
Resources for "Natural Language Processing" Coursera course.
Stars: ✭ 969 (+2836.36%)
Mutual labels:  jupyter-notebook
Pm Pyro
PyMC3-like Interface for Pyro
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Object detection tools
Object detection useful tools for TensorFlow Object Detection API
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Machinelearningdeeplearning
李宏毅2021机器学习深度学习笔记PPT作业
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
Gaze Estimation
A deep learning based gaze estimation framework implemented with PyTorch
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Simple Ssd For Beginners
This repository contains easy SSD(Single Shot MultiBox Detector) implemented with Pytorch and is easy to read and learn
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Madmom tutorials
Tutorials for the madmom package.
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
Geemap
A Python package for interactive mapping with Google Earth Engine, ipyleaflet, and folium
Stars: ✭ 959 (+2806.06%)
Mutual labels:  jupyter-notebook
Pyhat
Python Hyperspectral Analysis Tools
Stars: ✭ 32 (-3.03%)
Mutual labels:  jupyter-notebook
Voice emotion
Detecting emotion in voices
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Multitask Learning
MSc group project: Reproduction of 'Multi-Task Learning using Uncertainty to Weigh Losses for Scene Geometry and Semantics'; A. Kendall, Y. Gal, R. Cipolla
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook
Aws Deepracer Workshops
DeepRacer workshop content
Stars: ✭ 968 (+2833.33%)
Mutual labels:  jupyter-notebook
Pytorch Softplus Normalization Uncertainty Estimation Bayesian Cnn
PyTorch code for Paper "Uncertainty Estimations by Softplus normalization in Bayesian Convolutional Neural Networks with Variational Inference"
Stars: ✭ 33 (+0%)
Mutual labels:  jupyter-notebook

SANet-Keras

An unofficial implementation of SANet for crowd counting in Keras-TF.


Paper:

Results now:

On dataset ShanghaiTech B

Still far from the performance in the original paper(MAE 8.4)

MAE MSE MAPE Mean DM Distance
13.727 22.726 11.9 27.065

Dataset:

Training Parameters:

  1. Loss = ssim_loss + L2

  2. Optimizer = Adam(lr=1e-4)

  3. Data augmentation: Flip horizontally.

  4. Patch: No patch, input the whole image, output the same shape DM.

  5. Instance normalization: No IN layers at present, since network with IN layers is very hard to train and IN layers didn't show improvement to the network in my experiments.

  6. Output Zeros: The density map output may fade to zeros in 95%+ random initialization, I tried the initialization method in the original paper while it didn't work. In the past, when this happens, I just restarted the kernel and re-run. But now, I tried to train different modules(1-5) separately in the first several epochs to get relatively reasonable weights:

    structure_lite, and it worked out to greatly decrease the probability of the zero-output-phenomena. Any other question, welcome to contact me.

  7. Weights: Got best weights in 251-th epoch(300 epochs in total), and here is the loss records:

    Loss_records

  8. Prediction example:

    example

Run:

  1. Download dataset;
  2. Data generation: run thegenerate_datasets.ipynb.
  3. Run the main.ipynb to train the model and do the test.

Abstraction:

  1. Network = encoder + decoder, model plot is here:

    Network encoder decoder
    Composition scale aggregation module conv2dTranspose
    Usage extract multi-scale features generate high resolution density map
  2. Loss:

    Loss = loss_formular

  3. Normalization layer:

    • Ease the training process;
    • Reduce 'statistic shift problem'.
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].