All Projects → siddk → Deep Nlp

siddk / Deep Nlp

Tensorflow Tutorial files and Implementations of various Deep NLP and CV Models.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Deep Nlp

Tensorflow-Wide-Deep-Local-Prediction
This project demonstrates how to run and save predictions locally using exported tensorflow estimator model
Stars: ✭ 28 (-45.1%)
Mutual labels:  deep, tensorflow-tutorials
Tensorflow Mnist Tutorial
MNIST classification in Tensorflow using Django
Stars: ✭ 36 (-29.41%)
Mutual labels:  tensorflow-tutorials
Tensorlayer
Deep Learning and Reinforcement Learning Library for Scientists and Engineers 🔥
Stars: ✭ 6,796 (+13225.49%)
Mutual labels:  tensorflow-tutorials
My Very Deep Caffe
This is an implementation of very deep two stream CNNs for action recognition. The implementation is inspired by Wang et. al., https://github.com/yjxiong/caffe. Some improvements from Wang's implementation include reading videos from LDMB database, faster testing using LDMB interface. The aim is to work better with big dataset such as UCF101, HMDB51, Sports1M and ActivityNet easily.
Stars: ✭ 21 (-58.82%)
Mutual labels:  deep
Tensorflow Tutorial
Some interesting TensorFlow tutorials for beginners.
Stars: ✭ 893 (+1650.98%)
Mutual labels:  tensorflow-tutorials
Model Free Episodic Control
This is the implementation of paper Model Free Episodic Control
Stars: ✭ 31 (-39.22%)
Mutual labels:  deep
My Tensorflow Tutorials
This repo contains all of my TensorFlow tutorials
Stars: ✭ 795 (+1458.82%)
Mutual labels:  tensorflow-tutorials
Ludwig
Data-centric declarative deep learning framework
Stars: ✭ 8,018 (+15621.57%)
Mutual labels:  deep
Satellite imagery analysis
Implementation of different techniques to find insights from the satellite data using Python.
Stars: ✭ 31 (-39.22%)
Mutual labels:  deep
Tensorflow examples
reference code for tensorflow
Stars: ✭ 14 (-72.55%)
Mutual labels:  tensorflow-tutorials
Tensorflow Find Object
📸 A simple application to demonstrate TensorflowJS using mobile net model to predict objects via camera API.
Stars: ✭ 12 (-76.47%)
Mutual labels:  tensorflow-tutorials
Machine Learning Collection
A resource for learning about ML, DL, PyTorch and TensorFlow. Feedback always appreciated :)
Stars: ✭ 883 (+1631.37%)
Mutual labels:  tensorflow-tutorials
Letslearnai.github.io
Lets Learn AI
Stars: ✭ 33 (-35.29%)
Mutual labels:  tensorflow-tutorials
Seq2seq Signal Prediction
Signal forecasting with a Sequence-to-Sequence (seq2seq) Recurrent Neural Network (RNN) model in TensorFlow - Guillaume Chevalier
Stars: ✭ 890 (+1645.1%)
Mutual labels:  tensorflow-tutorials
Tensorflow Serving sidecar
Serve machine learning models using tensorflow serving
Stars: ✭ 41 (-19.61%)
Mutual labels:  tensorflow-tutorials
Variational Autoencoder
Variational autoencoder implemented in tensorflow and pytorch (including inverse autoregressive flow)
Stars: ✭ 807 (+1482.35%)
Mutual labels:  deep
Pytorch Forecasting
Time series forecasting with PyTorch
Stars: ✭ 849 (+1564.71%)
Mutual labels:  deep
Tensorflow In Practice Specialization
DeepLearning.AI TensorFlow Developer Professional Certificate Specialization
Stars: ✭ 29 (-43.14%)
Mutual labels:  tensorflow-tutorials
Tensorflow From Zero To One
TensorFlow 最佳学习资源大全(含课程、书籍、博客、公开课等内容)
Stars: ✭ 1,052 (+1962.75%)
Mutual labels:  tensorflow-tutorials
Deeptrading
Deep Neural Network Trading collection of Tensorflow Jupyter notebooks
Stars: ✭ 41 (-19.61%)
Mutual labels:  tensorflow-tutorials

Deep - NLP

This repository contains Tensorflow implementations of various deep learning models, with a focus on problems in Natural Language Processing. Each individual subdirectory is self-contained, addressing one specific model.

Models

The following models are implemented:

  • mnist_nn/: A simple one-layer classifier implemented as part of the Tensorflow Tutorial for the MNIST Handwriting Classification task. Very simple model, contained in a single python script, just to show off the Tensorflow basics.

  • mnist_cnn/: A three-layer Convolutional Neural Network for the MNIST Handwritten Digit Classification task.

  • langmod_nn/: A three-layer Feed-Forward Bigram Language model that tries to predict the next word in the corpus given the current word.

  • langmod_rnn/: A Recurrent Neural Network Language Model. Given a fixed-sized context window, predict the time-shifted (next) words in the sequence.

  • memory_networks/: Implementation of End-to-End Memory Networks: https://papers.nips.cc/paper/5846-end-to-end-memory-networks.pdf

  • seq2seq_mt/: Implementation of Encoder-Decoder Machine Translation for French-English Translation.

  • multilingual_embeddings/: Multi-Task Translation implementation, for the purpose of learning language-agnostic sentence embeddings (encoding multiple source languages in the same low-dimensional space).

  • cifar_cnn/: A multi-layer Convolutional Neural Network that follows the AlexNet convention for the CIFAR-10 image classification task. Given an image, classify it as an airplane, automobile, bird, cat, deer, dog, frog, horse, ship, or truck.

  • deep_q/: Implementation of Deep Q-Learning for Atari Pong.

  • variational_autoencoder/: Variational Autoencoder implementation, compresses and decodes MNIST digits, with
    demonstration of walking latent space.

Project Setup

There are a lot of ways to set up and train Tensorflow models, which can be very confusing. With the exception of the simple MNIST NN from the Tensorflow Tutorial, each of the above model subdirectories is broken up in the following way:

train.py: This file is the core file for training a given model (think of this as the main script for the entire directory). This script loads the data, performs any preprocessing to format the data for training and testing, and then builds and trains the model. Usually contains code to output training details to stdout, as well as code to save/serialize the model parameters periodically.

preprocessor/: Subdirectory that contains any data processing code (i.e. code to read raw data like text or images, and convert it to numerical form for training/testing).

model/:

  • model.py: Class definition for the model's neural network. Tensorflow at its core is a system for building symbolic computational graphs, and everything in Tensorflow is either expressed as a raw Tensor, or a Tensor operation. Because of this, building a model consists of building different graphs and operations to handle the inference of the model, how to evaluate the loss/cost, and how to perform training (usually via backpropagation). Because of this, each class definition consists of the following three functions:

    • inference: This is the crux of any neural network. This function is responsible for building all the layers of the network, from the input, all the way to the final layer, just before the loss is calculated.

    • loss: Using the output from the inference function, this function evaluates the loss used for training the model. For example, the loss function might take in the logits from the softmax layer of a classification model (say like in the MNIST model), and calculate the cross-entropy loss with the true labels of the input data.

    • train: The train function builds the training operation, given the cost calculated in the loss function. This function computes the gradients, and sets up the optimizer (i.e. SGD, Adam, Adagrad, etc.). Any learning rate decay is also performed during this step.

data/: A data subdirectory, for storing raw data.

log/: A log directory, consisting of two parts - summaries, and checkpoints. Each of the above Tensorflow models have code to store Tensorboard-formatted Summary files to track things like loss over time, accuracy, gradients, etc, and these logs are stored in logs/summaries. The models also have code to save and serialize all the parameters during training, allowing for easy restoration of interrupted training, or for just loading fully trained models.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].