All Projects → xilaili → Aognet

xilaili / Aognet

Licence: mit
Code for CVPR 2019 paper: " Learning Deep Compositional Grammatical Architectures for Visual Recognition"

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Aognet

Iresnet
Improved Residual Networks (https://arxiv.org/pdf/2004.04989.pdf)
Stars: ✭ 163 (+23.48%)
Mutual labels:  deep-neural-networks, image-classification, imagenet
Imgclsmob
Sandbox for training deep learning networks
Stars: ✭ 2,405 (+1721.97%)
Mutual labels:  image-classification, mxnet, imagenet
Neural Backed Decision Trees
Making decision trees competitive with neural networks on CIFAR10, CIFAR100, TinyImagenet200, Imagenet
Stars: ✭ 411 (+211.36%)
Mutual labels:  image-classification, imagenet, cifar10
Torchdistill
PyTorch-based modular, configuration-driven framework for knowledge distillation. 🏆18 methods including SOTA are implemented so far. 🎁 Trained models, training logs and configurations are available for ensuring the reproducibiliy.
Stars: ✭ 177 (+34.09%)
Mutual labels:  image-classification, imagenet, cifar10
Petridishnn
Code for the neural architecture search methods contained in the paper Efficient Forward Neural Architecture Search
Stars: ✭ 112 (-15.15%)
Mutual labels:  image-classification, imagenet, cifar10
BottleneckTransformers
Bottleneck Transformers for Visual Recognition
Stars: ✭ 231 (+75%)
Mutual labels:  imagenet, image-classification, cifar10
Image classification cifar 10
Image Classification on CIFAR-10 Dataset using Multi Layer Perceptrons in Python from Scratch.
Stars: ✭ 18 (-86.36%)
Mutual labels:  deep-neural-networks, image-classification, cifar10
One Pixel Attack Keras
Keras implementation of "One pixel attack for fooling deep neural networks" using differential evolution on Cifar10 and ImageNet
Stars: ✭ 1,097 (+731.06%)
Mutual labels:  imagenet, cifar10
Deep Ranking
Learning Fine-grained Image Similarity with Deep Ranking is a novel application of neural networks, where the authors use a new multi scale architecture combined with a triplet loss to create a neural network that is able to perform image search. This repository is a simplified implementation of the same
Stars: ✭ 64 (-51.52%)
Mutual labels:  deep-neural-networks, image-classification
Regnet
Pytorch implementation of network design paradigm described in the paper "Designing Network Design Spaces"
Stars: ✭ 129 (-2.27%)
Mutual labels:  image-classification, imagenet
Pytorch Classification
Classification with PyTorch.
Stars: ✭ 1,268 (+860.61%)
Mutual labels:  imagenet, cifar10
Imagenet
Trial on kaggle imagenet object localization by yolo v3 in google cloud
Stars: ✭ 56 (-57.58%)
Mutual labels:  image-classification, imagenet
Divide And Co Training
[Paper 2020] Towards Better Accuracy-efficiency Trade-offs: Divide and Co-training. Plus, an image classification toolbox includes ResNet, Wide-ResNet, ResNeXt, ResNeSt, ResNeXSt, SENet, Shake-Shake, DenseNet, PyramidNet, and EfficientNet.
Stars: ✭ 54 (-59.09%)
Mutual labels:  image-classification, imagenet
Gluon2pytorch
Gluon to PyTorch deep neural network model converter
Stars: ✭ 70 (-46.97%)
Mutual labels:  deep-neural-networks, mxnet
Cv Pretrained Model
A collection of computer vision pre-trained models.
Stars: ✭ 995 (+653.79%)
Mutual labels:  image-classification, mxnet
Tf Mobilenet V2
Mobilenet V2(Inverted Residual) Implementation & Trained Weights Using Tensorflow
Stars: ✭ 85 (-35.61%)
Mutual labels:  image-classification, imagenet
Sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on Apache MXNet
Stars: ✭ 990 (+650%)
Mutual labels:  deep-neural-networks, mxnet
Bsconv
Reference implementation for Blueprint Separable Convolutions (CVPR 2020)
Stars: ✭ 84 (-36.36%)
Mutual labels:  image-classification, cifar10
Labeld
LabelD is a quick and easy-to-use image annotation tool, built for academics, data scientists, and software engineers to enable single track or distributed image tagging. LabelD supports both localized, in-image (multi-)tagging, as well as image categorization.
Stars: ✭ 129 (-2.27%)
Mutual labels:  deep-neural-networks, image-classification
Mxnet Finetuner
An all-in-one Deep Learning toolkit for image classification to fine-tuning pretrained models using MXNet.
Stars: ✭ 100 (-24.24%)
Mutual labels:  deep-neural-networks, mxnet

The new version code and pretrained models are available at https://github.com/iVMCL/AOGNets, and more code and models will be released.

Learning Deep Compositional Grammatical Architectures for Visual Recognition

This repository contains the code (in MXNet) for our CVPR2019 paper: "Learning Deep Compositional Grammatical Architectures for Visual Recognition" paper by Xilai Li, Tianfu Wu*, Xi Song. (* Corresponding Author)

Citation

If you find our project useful in your research, please consider citing:

@article{li2017aognet,
  title={Learning Deep Compositional Grammatical Architectures for Visual Recognition},
  author={Xilai Li, Tianfu Wu, Xi Song, Hamid Krim},
  journal={arXiv preprint arXiv:1711.05847},
  year={2017}
}

Contents

  1. Introduction
  2. Contacts

Introduction

An AOGNet consists of a number of stages each of which is composed of a number of AOG building blocks. An AOG building block is designed based on a principled AND-OR grammar and represented by a hierarchical and compositional AND-OR graph. There are three types of nodes: an AND-node explores composition, whose input is computed by concatenating features of its child nodes; an OR-node represents alternative ways of composition in the spirit of exploitation, whose input is the element-wise sum of features of its child nodes; and a Terminal-node takes as input a channel-wise slice of the input feature map of the AOG building block. AOGNets aim to harness the best of two worlds (grammar models and deep neural networks) in representation learning with end-to-end training.

Contacts

email: [email protected]

Any discussions and contribution are welcomed!

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].