All Projects → openai → neural-gpu

openai / neural-gpu

Licence: other
Code for the Neural GPU model originally described in "Neural GPUs Learn Algorithms"

Programming Languages

python
139335 projects - #7 most used programming language

Labels

Projects that are alternatives of or similar to neural-gpu

Hotpur
A fork of Purpur that aims to improve performance and add FabricMC compatibility.
Stars: ✭ 17 (-83%)
Mutual labels:  paper
FSL-Mate
FSL-Mate: A collection of resources for few-shot learning (FSL).
Stars: ✭ 1,346 (+1246%)
Mutual labels:  paper
STACP
Joint Geographical and Temporal Modeling based on Matrix Factorization for Point-of-Interest Recommendation - ECIR 2020
Stars: ✭ 19 (-81%)
Mutual labels:  paper
Purpur
Purpur is a drop-in replacement for Paper servers designed for configurability, and new fun and exciting gameplay features.
Stars: ✭ 1,224 (+1124%)
Mutual labels:  paper
BookLibrary
Book Library of P&W Studio
Stars: ✭ 13 (-87%)
Mutual labels:  paper
papers
Summarizing the papers I have read (Japanese)
Stars: ✭ 38 (-62%)
Mutual labels:  paper
AutoInAgda
Proof automation – for Agda, in Agda.
Stars: ✭ 38 (-62%)
Mutual labels:  paper
best AI papers 2021
A curated list of the latest breakthroughs in AI (in 2021) by release date with a clear video explanation, link to a more in-depth article, and code.
Stars: ✭ 2,740 (+2640%)
Mutual labels:  paper
atari-demo
Code for the blog post "Learning Montezuma’s Revenge from a Single Demonstration"
Stars: ✭ 21 (-79%)
Mutual labels:  paper
NLP-Natural-Language-Processing
Projects and useful articles / links
Stars: ✭ 149 (+49%)
Mutual labels:  paper
Tensorflow-YOLACT
Implementation of the paper "YOLACT Real-time Instance Segmentation" in Tensorflow 2
Stars: ✭ 97 (-3%)
Mutual labels:  paper
modules
The official repository for our paper "Are Neural Nets Modular? Inspecting Functional Modularity Through Differentiable Weight Masks". We develop a method for analyzing emerging functional modularity in neural networks based on differentiable weight masks and use it to point out important issues in current-day neural networks.
Stars: ✭ 25 (-75%)
Mutual labels:  paper
3d-detection-papers
The papers in this list are about 3d detection, especially those using point clouds.
Stars: ✭ 77 (-23%)
Mutual labels:  paper
paper annotations
A place to keep track of all the annotated papers.
Stars: ✭ 96 (-4%)
Mutual labels:  paper
minie
An open information extraction system that provides compact extractions
Stars: ✭ 83 (-17%)
Mutual labels:  paper
ShapeTextureDebiasedTraining
Code and models for the paper Shape-Texture Debiased Neural Network Training (ICLR 2021)
Stars: ✭ 95 (-5%)
Mutual labels:  paper
ScriptBlockPlus
任意のブロックにスクリプトを追加するプラグインです。
Stars: ✭ 25 (-75%)
Mutual labels:  paper
bug-localization
Source code of the paper "Leveraging textual properties of bug reports to localize relevant source files".
Stars: ✭ 15 (-85%)
Mutual labels:  paper
material-appearance-similarity
Code for the paper "A Similarity Measure for Material Appearance" presented in SIGGRAPH 2019 and published in ACM Transactions on Graphics (TOG).
Stars: ✭ 22 (-78%)
Mutual labels:  paper
batu-gunting-kertas-nuxt
Rock Paper Scissors Game with Artificial Intellegence
Stars: ✭ 50 (-50%)
Mutual labels:  paper

Status: Archive (code is provided as-is, no updates expected)

Code for the Neural GPU model originally described in [[http://arxiv.org/abs/1511.08228]].

Running experiments

Running one instance

The following would use 256 filters to train on binary multiplication, then 4-ary, then decimal:

python neural_gpu_trainer.py --nmaps=256 --task=bmul,qmul,mul --progressive_curriculum=5

My typical invocation is something like

  CUDA_VISIBLE_DEVICES=0 python neural_gpu_trainer.py --random_seed=0 --max_steps=200000 --forward_max=201 --nmaps=256 --task=bmul,qmul,mul --time_till_eval=4 --progressive_curriculum=5 --train_dir=../logs/August-12-curriculum/forward_max=201-nmaps=256-task=bmul,qmul,mul-progressive_curriculum=5-random_seed=0

The tests on decimal carry were done using invocations like the following:

  CUDA_VISIBLE_DEVICES=0 neural_gpu_trainer.py --train_dir=../logs/run1 --random_seed=1 --max_steps=100000 --forward_max=201 --nmaps=128 --task=add --time_till_eval=4 --time_till_ckpt=1

You can find a list of options, and their default values, in neuralgpu/trainer.py.

Examining results

Loading and examining a model

examples/examples_for_loading_model.py gives a simple instance of loading a model and running it on an instance.

Plotting results

Something like python plots/get_pretty_score.py cachedlogs/*/*task=bmul,qmul,mul-* works. There are a lot of options to make it prettier (renaming stuff, removing some runs, changing titles, reordering, etc.). For example, one of my plots was made with

python get_pretty_score.py cachedlogs/A*/*256*[=,]mul-* --titles '256 filters|' --title 'Decimal multiplication is easier with curriculum' --task mul --remove_strings='|-progressive_curriculum=5' --exclude='layer|progressive' --order '4,2,1,3' --global-legend=1

Requirements

  • TensorFlow (see tensorflow.org for how to install)
  • Matplotlib for Python (sudo apt-get install python-matplotlib)
  • joblib

Credits

Original code by Lukasz Kaiser (lukaszkaiser). Modified by Eric Price (ecprice)

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].