All Projects → zackchase → Mxnet The Straight Dope

zackchase / Mxnet The Straight Dope

Licence: apache-2.0
An interactive book on deep learning. Much easy, so MXNet. Wow. [Straight Dope is growing up] ---> Much of this content has been incorporated into the new Dive into Deep Learning Book available at https://d2l.ai/.

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to Mxnet The Straight Dope

All Python Codes Of Ztm Course By Andrei Neagoie
Stars: ✭ 229 (-91.02%)
Mutual labels:  jupyter-notebook
Installations mac ubuntu windows
Installations for Data Science. Anaconda, RStudio, Spark, TensorFlow, AWS (Amazon Web Services).
Stars: ✭ 231 (-90.94%)
Mutual labels:  jupyter-notebook
Statannot
add statistical annotations (pvalue significance) on an existing boxplot generated by seaborn boxplot
Stars: ✭ 228 (-91.06%)
Mutual labels:  jupyter-notebook
Hamiltonian Nn
Code for our paper "Hamiltonian Neural Networks"
Stars: ✭ 229 (-91.02%)
Mutual labels:  jupyter-notebook
Dagmm
My attempt at reproducing the paper Deep Autoencoding Gaussian Mixture Model for Unsupervised Anomaly Detection
Stars: ✭ 231 (-90.94%)
Mutual labels:  jupyter-notebook
Introduction To Python
Python is an interpreted, high-level, general-purpose programming language. Created by Guido van Rossum and first released in 1991, Python's design philosophy emphasizes code readability with its notable use of significant white space. (This repository contains Python 3 Code)
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Quantiacs Python
Python version of Quantiacs toolbox and sample trading strategies
Stars: ✭ 230 (-90.98%)
Mutual labels:  jupyter-notebook
Awesome Pandas
A collection of resources for pandas (Python) and related subjects.
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Imaginary Numbers Are Real
Code To Accompany YouTube Series Imaginary Numbers Are Real
Stars: ✭ 231 (-90.94%)
Mutual labels:  jupyter-notebook
Wassdistance
Approximating Wasserstein distances with PyTorch
Stars: ✭ 229 (-91.02%)
Mutual labels:  jupyter-notebook
Nlp made easy
Explains nlp building blocks in a simple manner.
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Machine Learning By Andrew Ng In Python
Documenting my python implementation of Andrew Ng's Machine Learning course
Stars: ✭ 231 (-90.94%)
Mutual labels:  jupyter-notebook
Aotodata
朱小五写文章涉及到的数据分析,爬虫,源数据
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Stat451 Machine Learning Fs20
STAT 451: Intro to Machine Learning @ UW-Madison (Fall 2020)
Stars: ✭ 230 (-90.98%)
Mutual labels:  jupyter-notebook
Pystacknet
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Structuredinference
Structured Inference Networks for Nonlinear State Space Models
Stars: ✭ 230 (-90.98%)
Mutual labels:  jupyter-notebook
Stylegan2 Face Modificator
Simple Encoder, Generator and Face Modificator with StyleGAN2
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Relevant Search Book
Code and Examples for Relevant Search
Stars: ✭ 231 (-90.94%)
Mutual labels:  jupyter-notebook
Mattnet
MAttNet: Modular Attention Network for Referring Expression Comprehension
Stars: ✭ 232 (-90.91%)
Mutual labels:  jupyter-notebook
Learn Statistical Learning Method
Implementation of Statistical Learning Method, Second Edition.《统计学习方法》第二版,算法实现。
Stars: ✭ 228 (-91.06%)
Mutual labels:  jupyter-notebook

Deep Learning - The Straight Dope (Deprecated Please see d2l.ai)

This content has been moved to Dive into the Deep Learning Book freely available at https://d2l.ai/.

Abstract

This repo contains an incremental sequence of notebooks designed to teach deep learning, MXNet, and the gluon interface. Our goal is to leverage the strengths of Jupyter notebooks to present prose, graphics, equations, and code together in one place. If we're successful, the result will be a resource that could be simultaneously a book, course material, a prop for live tutorials, and a resource for plagiarising (with our blessing) useful code. To our knowledge there's no source out there that teaches either (1) the full breadth of concepts in modern deep learning or (2) interleaves an engaging textbook with runnable code. We'll find out by the end of this venture whether or not that void exists for a good reason.

Another unique aspect of this book is its authorship process. We are developing this resource fully in the public view and are making it available for free in its entirety. While the book has a few primary authors to set the tone and shape the content, we welcome contributions from the community and hope to coauthor chapters and entire sections with experts and community members. Already we've received contributions spanning typo corrections through full working examples.

Implementation with Apache MXNet

Throughout this book, we rely upon MXNet to teach core concepts, advanced topics, and a full complement of applications. MXNet is widely used in production environments owing to its strong reputation for speed. Now with gluon, MXNet's new imperative interface (alpha), doing research in MXNet is easy.

Dependencies

To run these notebooks, you'll want to build MXNet from source. Fortunately, this is easy (especially on Linux) if you follow these instructions. You'll also want to install Jupyter and use Python 3 (because it's 2017).

Slides

The authors (& others) are increasingly giving talks that are based on the content in this books. Some of these slide-decks (like the 6-hour KDD 2017) are gigantic so we're collecting them separately in this repo. Contribute there if you'd like to share tutorials or course material based on this books.

Translation

As we write the book, large stable sections are simultaneously being translated into 中文, available in a web version and via GitHub source.

Table of contents

Part 1: Deep Learning Fundamentals

Part 2: Applications

Part 3: Advanced Methods

Appendices

  • Appendix 1: Cheatsheets
    • Roadmap gluon
    • Roadmap PyTorch to MXNet (work in progress)
    • Roadmap Tensorflow to MXNet
    • Roadmap Keras to MXNet
    • Roadmap Math to MXNet

Choose your own adventure

We've designed these tutorials so that you can traverse the curriculum in more than one way.

  • Anarchist - Choose whatever you want to read, whenever you want to read it.
  • Imperialist - Proceed through all tutorials in order. In this fashion you will be exposed to each model first from scratch, writing all the code ourselves but for the basic linear algebra primitives and automatic differentiation.
  • Capitalist - If you don't care how things work (or already know) and just want to see working code in gluon, you can skip (from scratch!) tutorials and go straight to the production-like code using the high-level gluon front end.

Authors

This evolving creature is a collaborative effort (see contributors tab). The lead writers, assimilators, and coders include:

Inspiration

In creating these tutorials, we've have drawn inspiration from some the resources that allowed us to learn deep / machine learning with other libraries in the past. These include:

Contribute

  • Already, in the short time this project has been off the ground, we've gotten some helpful PRs from the community with pedagogical suggestions, typo corrections, and other useful fixes. If you're inclined, please contribute!
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].