All Projects → calico → Basenji

calico / Basenji

Licence: apache-2.0
Sequential regulatory activity predictions with deep convolutional neural networks.

Projects that are alternatives of or similar to Basenji

Fantasy Basketball
Scraping statistics, predicting NBA player performance with neural networks and boosting algorithms, and optimising lineups for Draft Kings with genetic algorithm. Capstone Project for Machine Learning Engineer Nanodegree by Udacity.
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Bug Hunting Colab
A Colab For Bug Hunting!
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Python4ds
Jupyter Notebooks used on my DataScience projects
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Deeplearningbookcode Volume1
Python/Jupyter notebooks for Volume 1 of "Deep Learning - From Basics to Practice" by Andrew Glassner
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Chess Alpha Zero
Chess reinforcement learning by AlphaGo Zero methods.
Stars: ✭ 1,868 (+1179.45%)
Mutual labels:  jupyter-notebook
Anomaly detection tuto
Anomaly detection tutorial on univariate time series with an auto-encoder
Stars: ✭ 144 (-1.37%)
Mutual labels:  jupyter-notebook
Face generator
DCGAN face generator 🧑.
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Deep Learning
Deep Learning Application Examples
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Deepschool.io
Deep Learning tutorials in jupyter notebooks.
Stars: ✭ 1,780 (+1119.18%)
Mutual labels:  jupyter-notebook
02456 Deep Learning With Pytorch
Exercises and supplementary material for the deep learning course 02456 using PyTorch.
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Segaware
Segmentation-Aware Convolutional Networks Using Local Attention Masks
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Dpca
An implementation of demixed Principal Component Analysis (a supervised linear dimensionality reduction technique)
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Bert binary text classification
Accompanying code for the Medium article
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Applied Dl 2018
Tel-Aviv Deep Learning Boot-camp: 12 Applied Deep Learning Labs
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Deep Learning Specialization Coursera
Deep Learning Specialization courses by Andrew Ng, deeplearning.ai
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Siamese Networks
Few Shot Learning by Siamese Networks, using Keras.
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Testbook
🧪 📗 Unit test your Jupyter Notebooks the right way
Stars: ✭ 146 (+0%)
Mutual labels:  jupyter-notebook
Multivariate Time Series Forecast Using Seq2seq In Tensorflow
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Keras Trigger Word
How to do Real Time Trigger Word Detection with Keras | DLology
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook
Deep Face Recognition
One-shot Learning and deep face recognition notebooks and workshop materials
Stars: ✭ 147 (+0.68%)
Mutual labels:  jupyter-notebook

Basenji

Sequential regulatory activity predictions with deep convolutional neural networks.

Basenji provides researchers with tools to:

  1. Train deep convolutional neural networks to predict regulatory activity along very long chromosome-scale DNA sequences
  2. Score variants according to their predicted influence on regulatory activity across the sequence and/or for specific genes.
  3. Annotate the distal regulatory elements that influence gene activity.
  4. Annotate the specific nucleotides that drive regulatory element function.

Akita

3D genome folding predictions with deep convolutional neural networks.

Akita provides researchers with tools to:

  1. Train deep convolutional neural networks to predict 2D contact maps along very long chromosome-scale DNA sequences
  2. Score variants according to their predicted influence on contact maps across the sequence and/or for specific genes.
  3. Annotate the specific nucleotides that drive genome folding.

Basset successor

This codebase offers numerous improvements and generalizations to its predecessor Basset, and I'll be using it for all of my ongoing work. Here are the salient changes.

  1. Basenji makes predictions in bins across the sequences you provide. You could replicate Basset's peak classification by simply providing smaller sequences and binning the target for the entire sequence.
  2. Basenji intends to predict quantitative signal using regression loss functions, rather than binary signal using classification loss functions.
  3. Basenji is built on TensorFlow, which offers myriad benefits, including distributed computing and a large and adaptive developer community.

However, this codebase is general enough to implement the Basset model, too. I have instructions for how to do that here.


Installation

Basenji/Akita were developed with Python3 and a variety of scientific computing dependencies, which you can see and install via requirements.txt for pip and environment.yml for Anaconda. For each case, we kept TensorFlow separate to allow you to choose the install method that works best for you. The codebase is compatible with the latest TensorFlow 2, but should also work with 1.15.

Run the following to install dependencies and Basenji with Anaconda.

    conda env create -f environment.yml
    conda install tensorflow (or tensorflow-gpu)
    python setup.py develop --no-deps

Alternatively, if you want to guarantee working versions of each dependency, you can install via a fully pre-specified environment.

    conda env create -f prespecified.yml
    conda install tensorflow (or tensorflow-gpu)
    python setup.py develop --no-deps

Or the following to install dependencies and Basenji with pip and setuptools.

    python setup.py develop
    pip install tensorflow (or tensorflow-gpu)

Then we recommend setting the following environmental variables.

  export BASENJIDIR=~/code/Basenji
  export PATH=$BASENJIDIR/bin:$PATH
  export PYTHONPATH=$BASENJIDIR/bin:$PYTHONPATH

To verify the install, launch python and run

    import basenji

Manuscripts

Models and (links to) data studied in various manuscripts are available in the manuscripts directory.


Documentation

At this stage, Basenji is something in between personal research code and accessible software for wide use. The primary challenge is uncertainty in what the best role for this type of toolkit is going to be in functional genomics and statistical genetics. The computational requirements don't make it easy either. Thus, this package is under active development, and I encourage anyone to get in touch to relate your experience and request clarifications or additional features, documentation, or tutorials.


Tutorials

These are a work in progress, so forgive incompleteness for the moment. If there's a task that you're interested in that I haven't included, feel free to post it as an Issue at the top.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].