All Projects → titu1994 → Neural Architecture Search

titu1994 / Neural Architecture Search

Licence: mit
Basic implementation of [Neural Architecture Search with Reinforcement Learning](https://arxiv.org/abs/1611.01578).

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Neural Architecture Search

BossNAS
(ICCV 2021) BossNAS: Exploring Hybrid CNN-transformers with Block-wisely Self-supervised Neural Architecture Search
Stars: ✭ 125 (-64.49%)
Mutual labels:  neural-architecture-search
regnet.pytorch
PyTorch-style and human-readable RegNet with a spectrum of pre-trained models
Stars: ✭ 50 (-85.8%)
Mutual labels:  neural-architecture-search
Autogluon
AutoGluon: AutoML for Text, Image, and Tabular Data
Stars: ✭ 3,920 (+1013.64%)
Mutual labels:  neural-architecture-search
TF-NAS
TF-NAS: Rethinking Three Search Freedoms of Latency-Constrained Differentiable Neural Architecture Search (ECCV2020)
Stars: ✭ 66 (-81.25%)
Mutual labels:  neural-architecture-search
ViPNAS
The official repo for CVPR2021——ViPNAS: Efficient Video Pose Estimation via Neural Architecture Search.
Stars: ✭ 32 (-90.91%)
Mutual labels:  neural-architecture-search
Deepswarm
Neural Architecture Search Powered by Swarm Intelligence 🐜
Stars: ✭ 263 (-25.28%)
Mutual labels:  neural-architecture-search
deep-learning-roadmap
my own deep learning mastery roadmap
Stars: ✭ 40 (-88.64%)
Mutual labels:  neural-architecture-search
Darts
Differentiable architecture search for convolutional and recurrent networks
Stars: ✭ 3,463 (+883.81%)
Mutual labels:  neural-architecture-search
Interstellar
Interstellar: Searching Recurrent Architecture for Knowledge Graph Embedding. NeurIPS 2020.
Stars: ✭ 28 (-92.05%)
Mutual labels:  neural-architecture-search
Awesome Automl Papers
A curated list of automated machine learning papers, articles, tutorials, slides and projects
Stars: ✭ 3,198 (+808.52%)
Mutual labels:  neural-architecture-search
rnn darts fastai
Implement Differentiable Architecture Search (DARTS) for RNN with fastai
Stars: ✭ 21 (-94.03%)
Mutual labels:  neural-architecture-search
nas-encodings
Encodings for neural architecture search
Stars: ✭ 29 (-91.76%)
Mutual labels:  neural-architecture-search
Archai
Reproducible Rapid Research for Neural Architecture Search (NAS)
Stars: ✭ 266 (-24.43%)
Mutual labels:  neural-architecture-search
mindware
An efficient open-source AutoML system for automating machine learning lifecycle, including feature engineering, neural architecture search, and hyper-parameter tuning.
Stars: ✭ 34 (-90.34%)
Mutual labels:  neural-architecture-search
Pnasnet.pytorch
PyTorch implementation of PNASNet-5 on ImageNet
Stars: ✭ 309 (-12.22%)
Mutual labels:  neural-architecture-search
ESNAC
Learnable Embedding Space for Efficient Neural Architecture Compression
Stars: ✭ 27 (-92.33%)
Mutual labels:  neural-architecture-search
NEATEST
NEATEST: Evolving Neural Networks Through Augmenting Topologies with Evolution Strategy Training
Stars: ✭ 13 (-96.31%)
Mutual labels:  neural-architecture-search
Adanet
Fast and flexible AutoML with learning guarantees.
Stars: ✭ 3,340 (+848.86%)
Mutual labels:  neural-architecture-search
Real Time Network
real-time network architecture for mobile devices and semantic segmentation
Stars: ✭ 308 (-12.5%)
Mutual labels:  neural-architecture-search
Randwirenn
Pytorch Implementation of: "Exploring Randomly Wired Neural Networks for Image Recognition"
Stars: ✭ 270 (-23.3%)
Mutual labels:  neural-architecture-search

Neural Architecture Search with Controller RNN

Basic implementation of Controller RNN from Neural Architecture Search with Reinforcement Learning and Learning Transferable Architectures for Scalable Image Recognition.

  • Uses Keras to define and train children / generated networks, which are defined in Tensorflow by the Controller RNN.
  • Define a state space by using StateSpace, a manager which adds states and handles communication between the Controller RNN and the user.
  • Controller manages the training and evaluation of the Controller RNN
  • NetworkManager handles the training and reward computation of a Keras model

Usage

At a high level : For full training details, please see train.py.

# construct a state space
state_space = StateSpace()

# add states
state_space.add_state(name='kernel', values=[1, 3])
state_space.add_state(name='filters', values=[16, 32, 64])

# create the managers
controller = Controller(tf_session, num_layers, state_space)
manager = NetworkManager(dataset, epochs=max_epochs, batchsize=batchsize)

# For number of trials
  sample_state = ...
  actions = controller.get_actions(sample_state)
  reward = manager.get_reward(actions)
  controller.train()

Implementation details

This is a very limited project.

  • Note: The controller eventually overfits to produce a single model always (generally a high performance model, but not always the highest). It seems there os a requirement to have a pool of controllers to avoid this overfitting, which would need distributed training. However, since the exploration rate is so high, there is a large set of very good models that are generated during the training process, and the controller yields a pretty strong model as well.
  • It doesnt have support for skip connections via 'anchor points' etc. (though it may not be that hard to implement it as a special state)
  • A lot of the details of the Controller were found in slides etc and not in the paper so I had to take many liberties when re-implementing it.
  • Learning rate, regularization strength, discount factor, exploration, beta value for EWA accuracy, clipping range etc are all random values (which make somewhat sense to me)
  • Single GPU model only. There would need to be a lot of modifications to this for multi GPU training (and I have just 1)

Implementation details were found from:

Result

I tried a toy CNN model with 4 CNN layers with different filter sizes (16, 32, 64) and kernel sizes (1, 3) to maximise score in 10 epochs of training on CIFAR-10.

After 50 steps, it converges to the "state space" of (3x3, 64)-(3x3, 64)-(3x3, 32)-(3x3, 64). Interestingly, this model performs very slightly better than a 4 x (3x3, 64) model, at least in the first 10 epochs.

Requirements

  • Keras >= 1.2.1
  • Tensorflow-gpu >= 1.2

Acknowledgements

Code heavily inspired by wallarm/nascell-automl

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].