All Projects → greydanus → Scribe

greydanus / Scribe

Realistic Handwriting with Tensorflow

Projects that are alternatives of or similar to Scribe

Stockpriceprediction
Stock Price Prediction using Machine Learning Techniques
Stars: ✭ 700 (+262.69%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Bitcoin Price Prediction Using Lstm
Bitcoin price Prediction ( Time Series ) using LSTM Recurrent neural network
Stars: ✭ 67 (-65.28%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Da Rnn
📃 **Unofficial** PyTorch Implementation of DA-RNN (arXiv:1704.02971)
Stars: ✭ 256 (+32.64%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Automatic Image Captioning
Generating Captions for images using Deep Learning
Stars: ✭ 84 (-56.48%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Image Caption Generator
[DEPRECATED] A Neural Network based generative model for captioning images using Tensorflow
Stars: ✭ 141 (-26.94%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Aulas
Aulas da Escola de Inteligência Artificial de São Paulo
Stars: ✭ 166 (-13.99%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Deep Learning Time Series
List of papers, code and experiments using deep learning for time series forecasting
Stars: ✭ 796 (+312.44%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Pytorch Learners Tutorial
PyTorch tutorial for learners
Stars: ✭ 97 (-49.74%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Understanding Pytorch Batching Lstm
Understanding and visualizing PyTorch Batching with LSTM
Stars: ✭ 125 (-35.23%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Rnn lstm from scratch
How to build RNNs and LSTMs from scratch with NumPy.
Stars: ✭ 156 (-19.17%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Lstm anomaly thesis
Anomaly detection for temporal data using LSTMs
Stars: ✭ 178 (-7.77%)
Mutual labels:  jupyter-notebook, lstm-neural-networks
Ibmquantumchallenge2020
Quantum Challenge problem sets
Stars: ✭ 193 (+0%)
Mutual labels:  jupyter-notebook
Trajectron Plus Plus
Code accompanying the ECCV 2020 paper "Trajectron++: Dynamically-Feasible Trajectory Forecasting With Heterogeneous Data" by Tim Salzmann*, Boris Ivanovic*, Punarjay Chakravarty, and Marco Pavone (* denotes equal contribution).
Stars: ✭ 191 (-1.04%)
Mutual labels:  jupyter-notebook
Cl Jupyter
An enhanced interactive Shell for Common Lisp (based on the Jupyter protocol)
Stars: ✭ 191 (-1.04%)
Mutual labels:  jupyter-notebook
Activitynet 2016 Cvprw
Tools to participate in the ActivityNet Challenge 2016 (NIPSW 2016)
Stars: ✭ 191 (-1.04%)
Mutual labels:  jupyter-notebook
Deep learning tutorial
[ko] 패스트캠퍼스 강의 자료
Stars: ✭ 193 (+0%)
Mutual labels:  jupyter-notebook
Py R Fcn Multigpu
Code for training py-faster-rcnn and py-R-FCN on multiple GPUs in caffe
Stars: ✭ 192 (-0.52%)
Mutual labels:  jupyter-notebook
Vanillacnn
Implementation of the Vanilla CNN described in the paper: Yue Wu and Tal Hassner, "Facial Landmark Detection with Tweaked Convolutional Neural Networks", arXiv preprint arXiv:1511.04031, 12 Nov. 2015. See project page for more information about this project. http://www.openu.ac.il/home/hassner/projects/tcnn_landmarks/ Written by Ishay Tubi : ishay2b [at] gmail [dot] com https://www.l
Stars: ✭ 191 (-1.04%)
Mutual labels:  jupyter-notebook
Pydata Cookbook
PyData Cookbook Project
Stars: ✭ 191 (-1.04%)
Mutual labels:  jupyter-notebook
Magic
MAGIC (Markov Affinity-based Graph Imputation of Cells), is a method for imputing missing values restoring structure of large biological datasets.
Stars: ✭ 189 (-2.07%)
Mutual labels:  jupyter-notebook

Scribe: Realistic Handwriting in Tensorflow

See blog post

Samples

"A project by Sam Greydanus" Sample output 1 "You know nothing Jon Snow" (print) Sample output 2 "You know nothing Jon Snow" (cursive) Sample output 3

"lowering the bias" Sample output 4 "makes the writing messier" Sample output 5 "but more random" Sample output 6

Jupyter Notebooks

For an easy intro to the code (along with equations and explanations) check out these Jupyter notebooks:

Getting started

  • install dependencies (see below).
  • download the repo
  • navigate to the repo in bash
  • download and unzip folder containing pretrained models: https://goo.gl/qbH2uw
    • place in this directory

Now you have two options:

  1. Run the sampler in bash: mkdir -p ./logs/figures && python run.py --sample --tsteps 700
  2. Open the sample.ipynb jupyter notebook and run cell-by-cell (it includes equations and text to explain how the model works)

About

This model is trained on the IAM handwriting dataset and was inspired by the model described by the famous 2014 Alex Graves paper. It consists of a three-layer recurrent neural network (LSTM cells) with a Gaussian Mixture Density Network (MDN) cap on top. I have also implemented the attention mechanism from the paper which allows the network to 'focus' on character at a time in a sequence as it draws them.

The model at one time step looks like this

Rolled model

Unrolling in time, we get Unrolled model

I've implemented the attention mechanism from the paper: Attention mechanism

Dependencies

  • All code is written in python 2.7. You will need:
  • Numpy
  • Matplotlib
  • TensorFlow 1.0
  • OPTIONAL: Jupyter (if you want to run sample.ipynb and dataloader.ipynb)
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].