All Projects → DOsinga → Deep_learning_cookbook

DOsinga / Deep_learning_cookbook

Licence: apache-2.0
Deep Learning Cookbox

Projects that are alternatives of or similar to Deep learning cookbook

Pandas Cookbook
Recipes for using Python's pandas library
Stars: ✭ 5,520 (+818.47%)
Mutual labels:  jupyter-notebook
Telemanom
A framework for using LSTMs to detect anomalies in multivariate time series data. Includes spacecraft anomaly data and experiments from the Mars Science Laboratory and SMAP missions.
Stars: ✭ 589 (-2%)
Mutual labels:  jupyter-notebook
Takehomedatachallenges
My solution to the book <A collection of Data Science Take-home Challenges>
Stars: ✭ 596 (-0.83%)
Mutual labels:  jupyter-notebook
Autodidact
A pedagogical implementation of Autograd
Stars: ✭ 585 (-2.66%)
Mutual labels:  jupyter-notebook
Tensorflow exercises
The codes I made while I practiced various TensorFlow examples
Stars: ✭ 588 (-2.16%)
Mutual labels:  jupyter-notebook
Python Deepdive
Python Deep Dive Course - Accompanying Materials
Stars: ✭ 590 (-1.83%)
Mutual labels:  jupyter-notebook
Dabl
Data Analysis Baseline Library
Stars: ✭ 585 (-2.66%)
Mutual labels:  jupyter-notebook
Courses
fast.ai Courses
Stars: ✭ 5,253 (+774.04%)
Mutual labels:  jupyter-notebook
Introduction To Python
Stars: ✭ 589 (-2%)
Mutual labels:  jupyter-notebook
Basic model scratch
Implementation of some classic Machine Learning model from scratch and benchmarking against popular ML library
Stars: ✭ 595 (-1%)
Mutual labels:  jupyter-notebook
Ml Interview
Preparing for machine learning interviews
Stars: ✭ 586 (-2.5%)
Mutual labels:  jupyter-notebook
Deep Learning For Hackers
Machine Learning tutorials with TensorFlow 2 and Keras in Python (Jupyter notebooks included) - (LSTMs, Hyperameter tuning, Data preprocessing, Bias-variance tradeoff, Anomaly Detection, Autoencoders, Time Series Forecasting, Object Detection, Sentiment Analysis, Intent Recognition with BERT)
Stars: ✭ 586 (-2.5%)
Mutual labels:  jupyter-notebook
Single Cell Tutorial
Single cell current best practices tutorial case study for the paper:Luecken and Theis, "Current best practices in single-cell RNA-seq analysis: a tutorial"
Stars: ✭ 594 (-1.16%)
Mutual labels:  jupyter-notebook
Grokking Deep Learning
this repository accompanies the book "Grokking Deep Learning"
Stars: ✭ 5,380 (+795.17%)
Mutual labels:  jupyter-notebook
Deeplearning Nlp
Introduction to Deep Learning for Natural Language Processing
Stars: ✭ 598 (-0.5%)
Mutual labels:  jupyter-notebook
Trtorch
PyTorch/TorchScript compiler for NVIDIA GPUs using TensorRT
Stars: ✭ 583 (-3%)
Mutual labels:  jupyter-notebook
Kobert
Korean BERT pre-trained cased (KoBERT)
Stars: ✭ 591 (-1.66%)
Mutual labels:  jupyter-notebook
Neuraltalk2
Efficient Image Captioning code in Torch, runs on GPU
Stars: ✭ 5,263 (+775.71%)
Mutual labels:  jupyter-notebook
Time Series Classification And Clustering
Time series classification and clustering code written in Python.
Stars: ✭ 599 (-0.33%)
Mutual labels:  jupyter-notebook
Yolov3 Complete Pruning
提供对YOLOv3及Tiny的多种剪枝版本,以适应不同的需求。
Stars: ✭ 596 (-0.83%)
Mutual labels:  jupyter-notebook

Deep Learning Cookbook Notebooks

This repository contains 35 python notebooks demonstrating most of the key machine learning techniques in Keras. The notebooks accompany the book Deep Learning Cookbook but work well on their own. A GPU is not required to run them, but on a mere CPU things will take quite a while.

Getting started

To get started, setup a virtual env, install the requirements and start the notebook server:

git clone https://github.com/DOsinga/deep_learning_cookbook.git
cd deep_learning_cookbook
python3 -m venv venv3
source venv3/bin/activate
pip install -r requirements.txt
jupyter notebook

The notebooks

03.1 Using pre trained word embeddings

In this notebook, we'll use a pretrained word embedding model (Word2Vec) to explore how word embeddings allow us to explore similarities between words and relationships between words. For example, find the capital of a country or the main products of a company. We'll finish with a demonstration of using t-SNE to plot high dimensional spaces on a 2D graph.

03.2 Domain specific ranking using word2vec cosine distance

Building on the previous recipe, we'll use the distances between the words to do domain specific rankings. Specifically we'll look at countries. First we create a small classifier to find all countries in the set of words, based on a small sample. We'll then use a similar approach to show relevance for specific words for countries. For example, since cricket is closer to India than to Germany, cricket is probably more relevant. We can plot this on a world map which lights up countries based on their relevance for specific words.

04.1 Collect movie data from Wikipedia

This notebook shows how to download a dump of the Wikipedia and parse it to extract structured data by using the category and template information. We'll use this to create a set of movies including rating data.

04.2 Build a recommender system based on outgoing Wikipedia links

Based on the structured data extracted in the previous notebook, we'll train a network that learns to predict a movie based on the outgoing links on the corresponding Wikipedia page. This creates embeddings for the movies. This in turn lets us recommend movies based on other movies - similar movies are next to each other in the embedding space.

05.1 Generating Text in the Style of an Example Text

We train an LSTM to write Shakespeare. We'll follow this up with one that generates Python code by training a similar LSTM on the Python system codebase. Visualizing what the network has learned shows us what the Python producing network is paying attention to as it produces or read Python code.

06.1 Question matching

In this notebook we train a network to learn how to match questions and answers from stackoverflow; this sort of indexing than allows us to find given a question what the most likely answer in a database is. We try a variety of approaches to improve upon the first not terribly great results.

07.1 Text Classification

This notebook shows eight different machine learning approaches to classify texts into a variety of sentiments. The first three are classical learners, followed by a number of deep learning models, character or word based and lstm vs cnn. The best approach is to combine all approaches in one model.

07.2 Emoji Suggestions

We start by harvesting a large set of tweets and we keep the ones that contain exactly one emoji (you can skip this step, a training set is included). We then train a number of deep models to use the tweet minus the emoji to predict the missing emoji. We end up effectively with a model that can find the best emoji for a given bit of text.

07.3 Tweet Embeddings

Some experimental code (not included in the book) to semantically index tweets such that tweets that are similar show up next to each other; effectively doing what Word2Vec does for words, but now for tweets.

08.1 Sequence to sequence mapping

08.2 Import Gutenberg

Small notebook demonstrating how to download books from the Gutenberg project. Tokenizes a set of book in preparation of the subword tokenizing in the next notebook.

09.1 Reusing a pretrained image recognition network

Quick notebook demonstrating how to load a pretrained network and apply it on an image of, well, what else? a cat. Shows how to normalize the image and decode the predictions.

09.2 Images as embeddings

In this notebook we use the Flickr API to fetch a feed of search results for the search term cat. By running each result through a pre-trained network we get vectors that project the images in a 'space'. The center of that space in some way represents the most cat image possible. By reranking on distance to that center we can weed out images that are less cat like. Effectively we can improve upon the Flickr search results without knowing the content!

09.3 Retraining

10.1 Building an inverse image search service

11.1 Detecting Multiple Images

Use the fact that imag classification networks extract features per larger square sub-image to detect multiple dogs and cats in the same image or at least to know where in the image you can find your cat or dog. The approach her is a lot simpler than what is the state of the art, but also a lot easier to follow, so a good way to get started.

12.1 Activation Optimization

12.2 Neural Style

13.1 Quick Draw Cat Autoencoder

13.2 Variational Autoencoder

13.5 Quick Draw Autoencoder

14.1 Importing icons

14.2 Icon Autoencoding

14.2 Variational Autoencoder Icons

14.3 Icon GAN

14.4 Icon RNN

15.1 Song Classification

15.2 Index Local MP3s

15.3 Spotify Playlists

15.4 Train a music recommender

16.1 Productionize Embeddings

16.2 Prepare Keras model for Tensorflow Serving

16.3 Prepare model for iOS

16.4 Simple Text Generation

Simple Seq2Seq

Simple sequence-to-sequence mapping demo. The notebook shows how to teach a network how to form plurals.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].