All Projects → JSeam2 → Neural Ordinary Differential Equations

JSeam2 / Neural Ordinary Differential Equations

Sample implementation of Neural Ordinary Differential Equations

Projects that are alternatives of or similar to Neural Ordinary Differential Equations

Skills
个人的技能树仓库,主要包含个人机器学习以及深度学习的笔记
Stars: ✭ 240 (-0.83%)
Mutual labels:  jupyter-notebook
Hands On Computer Vision With Tensorflow 2
Hands-On Computer Vision with TensorFlow 2, published by Packt
Stars: ✭ 241 (-0.41%)
Mutual labels:  jupyter-notebook
Retail Demo Store
AWS Retail Demo Store is a sample retail web application and workshop platform demonstrating how AWS infrastructure and services can be used to build compelling customer experiences for eCommerce, retail, and digital marketing use-cases
Stars: ✭ 238 (-1.65%)
Mutual labels:  jupyter-notebook
Deepreplay
Deep Replay - Generate visualizations as in my "Hyper-parameters in Action!" series!
Stars: ✭ 240 (-0.83%)
Mutual labels:  jupyter-notebook
Tacotron2
Tacotron 2 - PyTorch implementation with faster-than-realtime inference
Stars: ✭ 3,300 (+1263.64%)
Mutual labels:  jupyter-notebook
Numerical Linear Algebra V2
Jupyter Notebooks for Computational Linear Algebra course, taught summer 2018 in USF MSDS program
Stars: ✭ 241 (-0.41%)
Mutual labels:  jupyter-notebook
Aravec
AraVec is a pre-trained distributed word representation (word embedding) open source project which aims to provide the Arabic NLP research community with free to use and powerful word embedding models.
Stars: ✭ 239 (-1.24%)
Mutual labels:  jupyter-notebook
Deeppicar
Deep Learning Autonomous Car based on Raspberry Pi, SunFounder PiCar-V Kit, TensorFlow, and Google's EdgeTPU Co-Processor
Stars: ✭ 242 (+0%)
Mutual labels:  jupyter-notebook
Pytorch Sentiment Analysis
Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
Stars: ✭ 3,209 (+1226.03%)
Mutual labels:  jupyter-notebook
6.006 2015 Notebooks
Stars: ✭ 241 (-0.41%)
Mutual labels:  jupyter-notebook
Malaya
Natural Language Toolkit for bahasa Malaysia, https://malaya.readthedocs.io/
Stars: ✭ 239 (-1.24%)
Mutual labels:  jupyter-notebook
Stat Nlp Book
Interactive Lecture Notes, Slides and Exercises for Statistical NLP
Stars: ✭ 240 (-0.83%)
Mutual labels:  jupyter-notebook
The Elements Of Statistical Learning Notebooks
Jupyter notebooks for summarizing and reproducing the textbook "The Elements of Statistical Learning" 2/E by Hastie, Tibshirani, and Friedman
Stars: ✭ 241 (-0.41%)
Mutual labels:  jupyter-notebook
Datascience
Data Science in Julia course for JuliaAcademy.com, taught by Huda Nassar
Stars: ✭ 239 (-1.24%)
Mutual labels:  jupyter-notebook
Tslab
Interactive JavaScript and TypeScript programming with Jupyter
Stars: ✭ 240 (-0.83%)
Mutual labels:  jupyter-notebook
Iclr2021 Openreviewdata
Crawl & visualize ICLR papers and reviews.
Stars: ✭ 238 (-1.65%)
Mutual labels:  jupyter-notebook
Udemy derinogrenmeyegiris
Udemy Derin Öğrenmeye Giriş Kursunun Uygulamaları ve Daha Fazlası
Stars: ✭ 239 (-1.24%)
Mutual labels:  jupyter-notebook
Loss toolbox Pytorch
PyTorch Implementation of Focal Loss and Lovasz-Softmax Loss
Stars: ✭ 240 (-0.83%)
Mutual labels:  jupyter-notebook
Jetcam
Easy to use Python camera interface for NVIDIA Jetson
Stars: ✭ 242 (+0%)
Mutual labels:  jupyter-notebook
Deeptextures
Code to synthesise textures using convolutional neural networks as described in Gatys et al. 2015 (http://arxiv.org/abs/1505.07376)
Stars: ✭ 241 (-0.41%)
Mutual labels:  jupyter-notebook

Neural Ordinary Differential Equations

Overview and Summary

I try to implement the findings in the paper in this repo. Here's a summary of what I think is significant information.

Neural Ordinary Differential Equations introduces an interesting way of specifiying a neural network. Instead of treating the neural network as a sequence of discrete states, the approach parameterizes the derivative of the hidden state using a neural network. This parameterizing of the hidden state provides a continuous depth model provides a bunch of useful properties.

Useful Properties I Understand

  1. Memory Efficiency: The specifications give a constant memory costs wrt to depth

  2. Adaptive Computation: Adapt level of error for efficiency, useful for real-time or low-power applications

  3. Scalable and Invertible Normalizing Flows: Continuous transformation allows for easier computation of change of variables. The paper derives a new fclass of invertible density models that avoids the bottleneck of normalizing flows, allowing the model to be trained directly by max-likelihood.

  4. Scalable and invertible normalizing flows: The continuous transformation makes the change of variables formula easier to compute. This allows the constructuion of a new class of invertible density models that avoids previous bottlenecks.

  5. Continuous Time-series Models: Able to model time-series data that arrive at arbitrary times unlike RNNs.

What Are Normalizing Flows? I am confused.

I got stumped on this one for a while. My reaction irl the whole time -> ( ・◇・)?

If you're a noob like me you would probably get stumped too. Here's what I understand without the math.

Imagine if you were in a conference and someone asked the presenter something really difficult. Making the presenter go ( ・◇・)? The presenter doesn't want to cop out and look stupid so he/she tries to answer the question. The presenter then simplifies the question and answers that simpified question instead. Not satisfactory, but it should gets the point across. This is essentially the idea behind Variational Inference. The quantity that describes this is termed as the posterior distribution.

What happens if this explanation seemed to hand wavy or oversimplistic? We need to find a slightly more detailed way of explaning the same thing. How should we tune the complexity of the explanation? This method of tuning the complexity is the idea behind Normalizing Flows. By using these normalizing flows, we can apply a sequence of invertible transformation (we can back and forth and not lose information) to transform that simple posterior distribution (ie. the explanation) into something more complex that captures the idea we want to describe.

This paper essentially provides a continuous formulation of the normalizing flow concept. While an elegant concept, it can get hard to compute terms, like the change of variables. Using the continuous formulation, the paper offers a method of using the trace of the mapping function, which is more efficient. Check the maths in the paper for a clearer picture.

Useful links

  1. Quora explanation by Sam Wang I used his analogy
  2. Variational Inference by David M. Blei
  3. Variational Inference with Normalizing Flows Paper
  4. Normalizing Flows as suggested by zdhNarsil

Limitations

  1. Unstraightforward Minibatching: Though minibatching can still be achieved by concatenating the states of each batch elements together to form an ODE.
  2. Uniqueness: A unique solution only exists if the neural netowrk has finate weights and Lipshitz nonlinearities like tanh or relu.
  3. Reversibility: Forward trajectory of the network is invertible in principle, but numerical errors will emerge in forward ODE solver and reverse ODE solver (though this can be reduce at the cost of more computation). Information is lost due to multiple initial value mapping to the final state, this is expected to be a problem if the system dynamics encoded optimization-like, convergent dynamics.

Link To Arxiv Paper

https://arxiv.org/abs/1806.07366

TODOS

To look into JJFORD link to Arxiv Paper https://arxiv.org/abs/1810.01367

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].