All Projects → davidrosenberg → Mlcourse

davidrosenberg / Mlcourse

Machine learning course materials.

Projects that are alternatives of or similar to Mlcourse

Ml19 20a
CS 771A: Introduction to Machine Learning, IIT Kanpur, 2019-20-autumn offering
Stars: ✭ 115 (-70.36%)
Mutual labels:  jupyter-notebook, course-materials
2016learnpython
Python Teaching, Seminars for 2nd year students of School of Linguistics NRU HSE
Stars: ✭ 32 (-91.75%)
Mutual labels:  jupyter-notebook, course-materials
Practical rl
A course in reinforcement learning in the wild
Stars: ✭ 4,741 (+1121.91%)
Mutual labels:  jupyter-notebook, course-materials
Ppd599
USC urban data science course series with Python and Jupyter
Stars: ✭ 1,062 (+173.71%)
Mutual labels:  jupyter-notebook, course-materials
Course Computational Literary Analysis
Course materials for Introduction to Computational Literary Analysis, taught at UC Berkeley in Summer 2018, 2019, and 2020, and at Columbia University in Fall 2020.
Stars: ✭ 74 (-80.93%)
Mutual labels:  jupyter-notebook, course-materials
Storytelling With Data
Course materials for Dartmouth Course: Storytelling with Data (PSYC 81.09).
Stars: ✭ 59 (-84.79%)
Mutual labels:  jupyter-notebook, course-materials
Practical dl
DL course co-developed by YSDA, HSE and Skoltech
Stars: ✭ 1,006 (+159.28%)
Mutual labels:  jupyter-notebook, course-materials
Deeplearning2020
course materials for introduction to deep learning 2020
Stars: ✭ 90 (-76.8%)
Mutual labels:  jupyter-notebook, course-materials
Natural Language Processing Specialization
This repo contains my coursework, assignments, and Slides for Natural Language Processing Specialization by deeplearning.ai on Coursera
Stars: ✭ 151 (-61.08%)
Mutual labels:  jupyter-notebook, course-materials
Supervisely Tutorials
🌈 Tutorials for Supervise.ly
Stars: ✭ 385 (-0.77%)
Mutual labels:  jupyter-notebook
Computer Vision Basics With Python Keras And Opencv
Full tutorial of computer vision and machine learning basics with OpenCV and Keras in Python.
Stars: ✭ 387 (-0.26%)
Mutual labels:  jupyter-notebook
Transformers Tutorials
Github repo with tutorials to fine tune transformers for diff NLP tasks
Stars: ✭ 384 (-1.03%)
Mutual labels:  jupyter-notebook
Njucs Course Material From Yikaizhang
南京大学计算机系 课程资料 作业 代码 实验报告 NJU-CS 课程分享计划 🍚​
Stars: ✭ 383 (-1.29%)
Mutual labels:  jupyter-notebook
Data Science And Machine Learning From Scratch
Implements common data science methods and machine learning algorithms from scratch in python. Intuition and theory behind the algorithms is also discussed.
Stars: ✭ 387 (-0.26%)
Mutual labels:  jupyter-notebook
Latent 3d points
Auto-encoding & Generating 3D Point-Clouds.
Stars: ✭ 385 (-0.77%)
Mutual labels:  jupyter-notebook
Bayesian changepoint detection
Methods to get the probability of a changepoint in a time series.
Stars: ✭ 387 (-0.26%)
Mutual labels:  jupyter-notebook
Dawp
Jupyter Notebooks and code for Derivatives Analytics with Python (Wiley Finance) by Yves Hilpisch.
Stars: ✭ 383 (-1.29%)
Mutual labels:  jupyter-notebook
2019 Cs109a
https://harvard-iacs.github.io/2019-CS109A/
Stars: ✭ 384 (-1.03%)
Mutual labels:  jupyter-notebook
Production Data Science
Production Data Science: a workflow for collaborative data science aimed at production
Stars: ✭ 388 (+0%)
Mutual labels:  jupyter-notebook
Toyplot
Interactive plotting for Python.
Stars: ✭ 389 (+0.26%)
Mutual labels:  jupyter-notebook

Notable Changes Since 2018

  • Added a note on a moment-matching interpretation of fitting logistic regression and more general softmax-style linear conditional probability models.

Notable Changes from 2017FOML to 2018

  • Elaborated on the case against sparsity in the lecture on elastic net, to complement the reasons for sparsity on the slide Lasso Gives Feature Sparsity: So What?.
  • Added a note on conditional expectations, since many students find the notation confusing.
  • Added a note on the correlated features theorem for elastic net, which was basically a translation of Zou and Hastie's 2005 paper "Regularization and variable selection via the elastic net." into the notation of our class, dropping an unnecessary centering condition, and using a more standard definition of correlation.
  • Changes to EM Algorithm presentation: Added several diagrams (slides 10-14) to give the general idea of a variational method, and made explicit that the marginal log-likelihood is exactly the pointwise supremum over the variational lower bounds (slides 31 and 32)).
  • Treatment of the representer theorem is now well before any mention of kernels, and is described as an interesting consequence of basic linear algebra: "Look how the solution always lies in the subspace spanned by the data. That's interesting (and obvious with enough practice). We can now constrain our optimization problem to this subspace..."
  • The kernel methods lecture was rewritten to significantly reduce references to the feature map. When we're just talking about kernelization, it seems like unneeded extra notation.
  • Replaced the 1-hour crash course in Lagrangian duality with a 10-minute summary of Lagrangian duality, which I actually never presented and left as optional reading.
  • Added a brief note on Thompson sampling for Bernoulli Bandits as a fun application for our unit on Bayesian statistics.
  • Significant improvement of the programming problem for lasso regression in Homework #2.
  • New written and programming problems on logistic regression in Homework #5 (showing the equivalence of the ERM and the conditional probability model formulations, as well as implementing regularized logistic regression).
  • New homework on backpropagation Homework #7 (with Philipp Meerkamp and Pierre Garapon).

Notable Changes from 2017 to 2017FOML

Notable Changes from 2016 to 2017

  • New lecture on geometric approach to SVMs (Brett)
  • New lecture on principal component analysis (Brett)
  • Added slide on k-means++ (Brett)
  • Added slides on explicit feature vector for 1-dim RBF kernel
  • Created notebook to regenerate the buggy lasso/elastic net plots from Hastie's book (Vlad)
  • L2 constraint for linear models gives Lipschitz continuity of prediction function (Thanks to Brian Dalessandro for pointing this out to me).
  • Expanded discussion of L1/L2/ElasticNet with correlated random variables (Thanks Brett for the figures)

Notable Changes from 2015 to 2016

Possible Future Topics

Basic Techniques

  • Gaussian processes
  • MCMC (or at least Gibbs sampling)
  • Importance sampling
  • Density ratio estimation (for covariate shift, anomaly detection, conditional probability modeling)
  • Local methods (knn, locally weighted regression, etc.)

Applications

  • Collaborative filtering / matrix factorization (building on this lecture on matrix factorization and Brett's lecture on PCA)
  • Learning to rank and associated concepts
  • Bandits / learning from logged data?
  • Generalized additive models for interpretable nonlinear fits (smoothing way, basis function way, and gradient boosting way)
  • Automated hyperparameter search (with GPs, random, hyperband,...)
  • Active learning
  • Domain shift / covariate shift adaptation
  • Reinforcement learning (minimal path to REINFORCE)

Latent Variable Models

  • PPCA / Factor Analysis and non-Gaussian generalizations
    • Personality types as example of factor analysis if we can get data?
  • Variational Autoencoders
  • Latent Dirichlet Allocation / topic models
  • Generative models for images and text (where we care about the human-perceived quality of what's generated rather than the likelihood given to test examples) (GANs and friends)

Bayesian Models

  • Relevance vector machines
  • BART
  • Gaussian process regression and conditional probability models

Technical Points

Other

  • Class imbalance
  • Black box feature importance measures (building on Ben's 2018 lecture)
  • Quantile regression and conditional prediction intervals (perhaps integrated into homework on loss functions);
  • More depth on basic neural networks: weight initialization, vanishing / exploding gradient, possibly batch normalization
  • Finish up 'structured prediction' with beam search / Viterbi
    • give probabilistic analogue with MEMM's/CRF's
  • Generative vs discriminative (Jordan & Ng's naive bayes vs logistic regression, plus new experiments including regularization)
  • Something about causality?
  • DART
  • LightGBM and CatBoost efficient handling of categorical features (i.e. handling categorical features in regression trees )

Citation Information

Creative Commons License
Machine Learning Course Materials by Various Authors is licensed under a Creative Commons Attribution 4.0 International License. The author of each document in this repository is considered the license holder for that document.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].