All Projects → Hulalazz → A-_Guide_-to_Data_Sciecne_from_mathematics

Hulalazz / A-_Guide_-to_Data_Sciecne_from_mathematics

Licence: other
It is a blueprint to data science from the mathematics to algorithms. It is not completed.

Programming Languages

HTML
75241 projects
TeX
3793 projects

Projects that are alternatives of or similar to A- Guide -to Data Sciecne from mathematics

ProximalOperators.jl
Proximal operators for nonsmooth optimization in Julia
Stars: ✭ 119 (+376%)
Mutual labels:  admm, numerical-optimization
NMFADMM
A sparsity aware implementation of "Alternating Direction Method of Multipliers for Non-Negative Matrix Factorization with the Beta-Divergence" (ICASSP 2014).
Stars: ✭ 39 (+56%)
Mutual labels:  admm
PyGRANSO
PyGRANSO: A PyTorch-enabled port of GRANSO with auto-differentiation
Stars: ✭ 19 (-24%)
Mutual labels:  numerical-optimization
pikaia
Modern Fortran Edition of the Pikaia Genetic Algorithm
Stars: ✭ 29 (+16%)
Mutual labels:  numerical-optimization
kGCN
A graph-based deep learning framework for life science
Stars: ✭ 91 (+264%)
Mutual labels:  gcn
allie
🤖 A machine learning framework for audio, text, image, video, or .CSV files (50+ featurizers and 15+ model trainers).
Stars: ✭ 93 (+272%)
Mutual labels:  model-compression
ANTsR
Advanced Normalization Tools in R
Stars: ✭ 101 (+304%)
Mutual labels:  statistical-learning
SBR
⌛ Introducing Self-Attention to Target Attentive Graph Neural Networks (AISP '22)
Stars: ✭ 22 (-12%)
Mutual labels:  recommender-systems
st-gcn-sl
Spatial Temporal Graph Convolutional Networks for Sign Language (ST-GCN-SL) Recognition
Stars: ✭ 18 (-28%)
Mutual labels:  gcn
LibAUC
An End-to-End Machine Learning Library to Optimize AUC (AUROC, AUPRC).
Stars: ✭ 115 (+360%)
Mutual labels:  gcn
Literatures-on-GNN-Acceleration
A reading list for deep graph learning acceleration.
Stars: ✭ 50 (+100%)
Mutual labels:  gcn
CAE-ADMM
CAE-ADMM: Implicit Bitrate Optimization via ADMM-Based Pruning in Compressive Autoencoders
Stars: ✭ 34 (+36%)
Mutual labels:  admm
Yolov5-distillation-train-inference
Yolov5 distillation training | Yolov5知识蒸馏训练,支持训练自己的数据
Stars: ✭ 84 (+236%)
Mutual labels:  model-compression
recsys2019
The complete code and notebooks used for the ACM Recommender Systems Challenge 2019
Stars: ✭ 26 (+4%)
Mutual labels:  recommender-systems
DLCV2018SPRING
Deep Learning for Computer Vision (CommE 5052) in NTU
Stars: ✭ 38 (+52%)
Mutual labels:  model-compression
JCLAL
JCLAL is a general purpose framework developed in Java for Active Learning.
Stars: ✭ 22 (-12%)
Mutual labels:  statistical-learning
Regularization-Pruning
[ICLR'21] PyTorch code for our paper "Neural Pruning via Growing Regularization"
Stars: ✭ 44 (+76%)
Mutual labels:  model-compression
Statistical Learning Method by-Dr.-Li-Hang
基于Julia 1.0写李航博士《统计学习方法》
Stars: ✭ 30 (+20%)
Mutual labels:  statistical-learning
mvGAE
Drug Similarity Integration Through Attentive Multi-view Graph Auto-Encoders (IJCAI 2018)
Stars: ✭ 27 (+8%)
Mutual labels:  gcn
EiCOS
A C++ Second Order Cone Solver based on Eigen
Stars: ✭ 12 (-52%)
Mutual labels:  numerical-optimization

A Guide to Data Sciecne: from Mathematics to Algorithms

It is a blueprint to data science from the mathematics to algorithms. It is not completed. It is of my own interest. It is web-source based. And some material draw from the original owners' blogs. Last but not least, thanks to the teachers who guide me to mathematics. And the copyright of pictures are not permmited by the original oweners. If this material infringes your copyright, you could contact me via Github and I will delete the pictures.

I hope that it is on the ADEPT way like Kalid Azad

ADEPT Method for Learning
Analogy Tell me what it is like
Diagram Help me visualize it
Example Allow me to experience it
Plain English Describe it with everyday word
Technical Definition Discuss the formal details

It is expected a part of A Guide to Data Science, where the overview of data science is roughly presented. It is used to attract the people to the world of data.

It includes the archicture, optimization methods and regularization, acceleration and compression of deep neural network. The state of the art is not discussed.

It includes not limited the following content:

  • basic introduction to probability and statistics;
  • sampling algorithm based on uniformly distributed data;
  • MCMC and stochastic methods;
  • generalized linear model and regression analysis;
  • basic machine learning;
  • numerical optimmization methods;
  • some materials on artifical neural network and deep learning;
  • probabilistic programming and graph models;
  • other data anslysis such as topological data analysis;
  • applications or models in practice such as recommender system, information retrieval, computer vision;
  • other computational intelligence such as simulated annealing.

As the name shown, it is a guide thus it does not cover all the methods or technques in data science or data mining. Deeper advanced topics such as probability correct algorithms are not discussed. Until now, it is web resource driven. There are many links on each topic but no concrete examples and codes.

The basic idea is machine learning = representation+evaluation + optimization. I would like formulate every machine learning problem into numerical optimization problems. The section Numerical optimmization may be too theoretical for practioner and too simple for the research on optimization.

It is draft and notes when I learn data science. It is planned to be a open source e-book.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].