All Projects → timsainb → ParametricUMAP_paper

timsainb / ParametricUMAP_paper

Licence: MIT license
Parametric UMAP embeddings for representation and semisupervised learning. From the paper "Parametric UMAP: learning embeddings with deep neural networks for representation and semi-supervised learning" (Sainburg, McInnes, Gentner, 2020).

Programming Languages

Jupyter Notebook
11667 projects
python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to ParametricUMAP paper

M-NMF
An implementation of "Community Preserving Network Embedding" (AAAI 2017)
Stars: ✭ 119 (-9.85%)
Mutual labels:  representation-learning, semisupervised-learning
dbMAP
A fast, accurate, and modularized dimensionality reduction approach based on diffusion harmonics and graph layouts. Escalates to millions of samples on a personal laptop. Adds high-dimensional big data intrinsic structure to your clustering and data visualization workflow.
Stars: ✭ 39 (-70.45%)
Mutual labels:  dimensionality-reduction, umap
ReductionWrappers
R wrappers to connect Python dimensional reduction tools and single cell data objects (Seurat, SingleCellExperiment, etc...)
Stars: ✭ 31 (-76.52%)
Mutual labels:  dimensionality-reduction, umap
Unsupervised-Learning-in-R
Workshop (6 hours): Clustering (Hdbscan, LCA, Hopach), dimension reduction (UMAP, GLRM), and anomaly detection (isolation forests).
Stars: ✭ 34 (-74.24%)
Mutual labels:  dimensionality-reduction, umap
UMAP.jl
Uniform Manifold Approximation and Projection (UMAP) implementation in Julia
Stars: ✭ 93 (-29.55%)
Mutual labels:  dimensionality-reduction, umap
Umap
Uniform Manifold Approximation and Projection
Stars: ✭ 5,268 (+3890.91%)
Mutual labels:  dimensionality-reduction, umap
partition
A fast and flexible framework for data reduction in R
Stars: ✭ 33 (-75%)
Mutual labels:  dimensionality-reduction
REGAL
Representation learning-based graph alignment based on implicit matrix factorization and structural embeddings
Stars: ✭ 78 (-40.91%)
Mutual labels:  representation-learning
FUSION
PyTorch code for NeurIPSW 2020 paper (4th Workshop on Meta-Learning) "Few-Shot Unsupervised Continual Learning through Meta-Examples"
Stars: ✭ 18 (-86.36%)
Mutual labels:  representation-learning
mathematics-statistics-for-data-science
Mathematical & Statistical topics to perform statistical analysis and tests; Linear Regression, Probability Theory, Monte Carlo Simulation, Statistical Sampling, Bootstrapping, Dimensionality reduction techniques (PCA, FA, CCA), Imputation techniques, Statistical Tests (Kolmogorov Smirnov), Robust Estimators (FastMCD) and more in Python and R.
Stars: ✭ 56 (-57.58%)
Mutual labels:  dimensionality-reduction
moses
Streaming, Memory-Limited, r-truncated SVD Revisited!
Stars: ✭ 19 (-85.61%)
Mutual labels:  dimensionality-reduction
GLOM-TensorFlow
An attempt at the implementation of GLOM, Geoffrey Hinton's paper for emergent part-whole hierarchies from data
Stars: ✭ 32 (-75.76%)
Mutual labels:  representation-learning
federated pca
Federated Principal Component Analysis Revisited!
Stars: ✭ 30 (-77.27%)
Mutual labels:  dimensionality-reduction
sef
A Python Library for Similarity-based Dimensionality Reduction
Stars: ✭ 24 (-81.82%)
Mutual labels:  dimensionality-reduction
Learning-From-Rules
Implementation of experiments in paper "Learning from Rules Generalizing Labeled Exemplars" to appear in ICLR2020 (https://openreview.net/forum?id=SkeuexBtDr)
Stars: ✭ 46 (-65.15%)
Mutual labels:  representation-learning
awesome-graph-self-supervised-learning
Awesome Graph Self-Supervised Learning
Stars: ✭ 805 (+509.85%)
Mutual labels:  representation-learning
PCC-pytorch
A pytorch implementation of the paper "Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control"
Stars: ✭ 57 (-56.82%)
Mutual labels:  representation-learning
causal-ml
Must-read papers and resources related to causal inference and machine (deep) learning
Stars: ✭ 387 (+193.18%)
Mutual labels:  representation-learning
reprieve
A library for evaluating representations.
Stars: ✭ 68 (-48.48%)
Mutual labels:  representation-learning
A-Barebones-Image-Retrieval-System
This project presents a simple framework to retrieve images similar to a query image.
Stars: ✭ 25 (-81.06%)
Mutual labels:  representation-learning

Open Example In Colab

Parametric UMAP (2020; Code for paper)

parametric-umap-algorithm

This repository contains the code needed to reproduce the results in the paper "Parametric UMAP: learning embeddings with deep neural networks for representation and semi-supervised learning" by Sainburg, McInnes, and Gentner (2020).

Citation:

@article{parametricumap,
  title={Parametric UMAP: learning embeddings with deep neural networks for representation and semi-supervised learning},
  author={Sainburg, Tim and McInnes, Leland and Gentner, Timothy Q},
}

How to use

The main implementation of this code is available in umap.parametric_umap in the UMAP repository (v0.5+). Most people reading this will want to use that code, and can ignore this repository.

The code in this repository is the 'messy' version. It has custom training loops which are a bit more verbose and customizable. It might be more useful for integrating UMAP into your custom models.

The code can be installed with python setup.py develop. Though, unless you're just trying to reproduce our results, you'll probably just want to pick through the notebooks and tfumap folder for the code relevant to your project.

In addition, we have a more verbose Colab notebook to walk you through the algorithm:

Parametric UMAP (verbose) Open In Colab

What's inside

This repo contains the code needed to produce all of the results in the paper. The network architectures we implement (in Tensorflow) are non-parametric UMAP, Parametric UMAP, a UMAP/AE hybrid, and a UMAP/classifier network hybrid.

network-outlines

The UMAP/classifier hybrid can be used for semisupervised learning on structured data. An example with the moons dataset is shown below, where in the left panel, the colored points are labeled training data, the grey points are unlabled data, and the background is the network's decision boundary.

semisupervised-example

The experiments inside use the following datasets and algorithms:

datasets


Project based on the cookiecutter data science project template. #cookiecutterdatascience

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].