All Projects → JizhiziLi → Animal Matting

JizhiziLi / Animal Matting

Github repository for the paper End-to-end Animal Image Matting

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Animal Matting

3d Pointcloud
Papers and Datasets about Point Cloud.
Stars: ✭ 179 (-50.69%)
Mutual labels:  datasets, segmentation
Multi object datasets
Multi-object image datasets with ground-truth segmentation masks and generative factors.
Stars: ✭ 121 (-66.67%)
Mutual labels:  datasets, segmentation
Awesome Iccv
ICCV2019最新录用情况
Stars: ✭ 305 (-15.98%)
Mutual labels:  segmentation
Inaspeechsegmenter
CNN-based audio segmentation toolkit. Allows to detect speech, music and speaker gender. Has been designed for large scale gender equality studies based on speech time per gender.
Stars: ✭ 352 (-3.03%)
Mutual labels:  segmentation
Deepcut
A Thai word tokenization library using Deep Neural Network
Stars: ✭ 330 (-9.09%)
Mutual labels:  segmentation
Tianchi Medical Lungtumordetect
天池医疗AI大赛[第一季]:肺部结节智能诊断 UNet/VGG/Inception/ResNet/DenseNet
Stars: ✭ 314 (-13.5%)
Mutual labels:  segmentation
Swift Composable Architecture
A library for building applications in a consistent and understandable way, with composition, testing, and ergonomics in mind.
Stars: ✭ 5,199 (+1332.23%)
Mutual labels:  composition
Medical Datasets
tracking medical datasets, with a focus on medical imaging
Stars: ✭ 296 (-18.46%)
Mutual labels:  datasets
Dr.sure
🏫DeepLearning学习笔记以及Tensorflow、Pytorch的使用心得笔记。Dr. Sure会不定时往项目中添加他看到的最新的技术,欢迎批评指正。
Stars: ✭ 365 (+0.55%)
Mutual labels:  datasets
Chakin
Simple downloader for pre-trained word vectors
Stars: ✭ 323 (-11.02%)
Mutual labels:  datasets
Scriptum
A fool's scriptum on functional programming
Stars: ✭ 346 (-4.68%)
Mutual labels:  composition
Akshare
AKShare is an elegant and simple financial data interface library for Python, built for human beings! 开源财经数据接口库
Stars: ✭ 4,334 (+1093.94%)
Mutual labels:  datasets
Lambda Talk
A Flock of Functions: Combinators, Lambda Calculus, & Church Encodings in JS
Stars: ✭ 315 (-13.22%)
Mutual labels:  composition
Unet
unet for image segmentation
Stars: ✭ 3,751 (+933.33%)
Mutual labels:  segmentation
Tailwindcss Classnames
TypeScript support for TailwindCSS
Stars: ✭ 305 (-15.98%)
Mutual labels:  composition
Clean Code Dotnet
🛁 Clean Code concepts and tools adapted for .NET
Stars: ✭ 4,425 (+1119.01%)
Mutual labels:  composition
Pointnet
PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation
Stars: ✭ 3,517 (+868.87%)
Mutual labels:  segmentation
Awesome Segmentation Saliency Dataset
A collection of some datasets for segmentation / saliency detection. Welcome to PR...😄
Stars: ✭ 315 (-13.22%)
Mutual labels:  datasets
Compoxure
Proxy middleware for express that enables composition of microservices.
Stars: ✭ 332 (-8.54%)
Mutual labels:  composition
Instaboost
Code for ICCV2019 paper "InstaBoost: Boosting Instance Segmentation Via Probability Map Guided Copy-Pasting"
Stars: ✭ 368 (+1.38%)
Mutual labels:  segmentation

End-to-end Animal Image Matting

This is the official repository of the paper End-to-end Animal Image Matting.

Google Colab Demo | Introduction | GFM | AM-2k | BG-20k | Results Demo | Installation | Inference Code | Statement


🚀 News

The training code and the dataset AM-2k will be released after paper review.

[2020-11-17]: Create Google Colab demo to benefit users who want to have a try online.

[2020-11-06]: Release the dataset BG-20k. Please fill out this agreement and send it to [email protected] from your academic email address to request.

[2020-11-03]: Publish the inference code and a pretrained model that can be used to test on your own animal images.

[2020-10-27]: Publish a video demo contains motivation, network, datasets, and test results on an animal video.

Demo on Google Colab

For those who do not have GPUs in their environment or only want to have a simple try online, you can try our Google Colab demo to generate the results for your images easily.

Introduction

This repository contains the code, datasets, models, test results and a video demo for the paper End-to-end Animal Image Matting. We propose a novel Glance and Focus Matting network (GFM), which employs a shared encoder and two separate decoders to learn both tasks in a collaborative manner for end-to-end animal matting. We also establish a novel Animal Matting dataset (AM-2k) to serve for end-to-end matting task. Furthermore, we investigate the domain gap issue between composition images and natural images systematically, propose a carefully designed composite route RSSN and a large-scale high-resolution background dataset (BG-20k) to serve as better candidates for composition.

Here is a video demo to illustrate the motivation, the network, the datasets, and the test results on an animal video.

We have released the inference code, a pretrained model and the Google Colab demo, which can be found in section inference code for more details. We have also published dataset BG-20k, please follow the guidance in section Bg-20k to access. Since the paper is currently under review, the dataset AM-2k, training code and the rest pretrained models will be made public after review.

GFM

The architecture of our proposed end-to-end method GFM is illustrated below. We adopt three kinds of Representation of Semantic and Transition Area (RoSTa) -TT, -FT, -BT within our method.

We trained GFM with three backbones, -(d) (DenseNet-121), -(r) (ResNet-34), and -(r2b) (ResNet-34 with 2 extra blocks). The trained model for each backbone can be downloaded via the link listed below.

GFM(d)-TT GFM(r)-TT GFM(r2b)-TT
coming soon coming soon model

AM-2k

Our proposed AM-2k contains 2,000 high-resolution natural animal images from 20 categories along with manually labeled alpha mattes. Some examples are shown as below, more can be viewed in the video demo.

BG-20k

Our proposed BG-20k contains 20,000 high-resolution background images excluded salient objects, which can be used to help generate high quality synthetic data. Some examples are shown as below, more can be viewed in the video demo.

The BG-20k dataset is publish now!!

You can request it by filling out this agreement and sending it to [email protected] from your academic email address. Please note the dataset can be only used for research purpose.

Results Demo

We test GFM on our AM-2k test dataset and show the results as below. More results on AM-2k test set can be found here.

Installation

Requirements:

  • Python 3.6.5+ with Numpy and scikit-image
  • Pytorch (version 1.4.0)
  • Torchvision (version 0.5.0)
  1. Clone this repository

    git clone https://github.com/JizhiziLi/animal-matting.git

  2. Go into the repository

    cd animal-matting

  3. Create conda environment and activate

    conda create -n animalmatting python=3.6.5

    conda activate animalmatting

  4. Install dependencies, install pytorch and torchvision separately if you need

    pip install -r requirements.txt

    conda install pytorch==1.4.0 torchvision==0.5.0 cudatoolkit=10.1 -c pytorch

Our code has been tested with Python 3.6.5, Pytorch 1.4.0, Torchvision 0.5.0, CUDA 10.1 on Ubuntu 18.04.

Inference Code - How to Test on Your Images

Here we provide the procedure of testing on sample images by our pretrained models:

  1. Download pretrained models as shown in section GFM, unzip to folder models/

  2. Save your high-resolution sample images in folder samples/original/.

  3. Setup parameters in scripts/deploy_samples.sh and run it

    chmod +x scripts/*

    ./scripts/deploy_samples.sh

  4. The results of alpha matte and transparent color image will be saved in folder samples/result_alpha/. and samples/result_color/.

We show some sample images from the internet, the predicted alpha mattes, and their transparent results as below. (We adopt arch='e2e_resnet34_2b_gfm_tt' and use hybrid testing strategy.)

Statement

This project is for research purpose only, please contact us for the licence of commercial use. For any other questions please contact [email protected].

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].