All Projects → NCBI-Hackathons → Spew

NCBI-Hackathons / Spew

Licence: mit
Automatic Packaging and Distribution of Bioinformatics Pipelines

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Spew

Argo Cd
Declarative continuous deployment for Kubernetes.
Stars: ✭ 7,887 (+37457.14%)
Mutual labels:  pipeline
Scipipe
Robust, flexible and resource-efficient pipelines using Go and the commandline
Stars: ✭ 826 (+3833.33%)
Mutual labels:  pipeline
Quagmir
A python-based isomiR quantification and analysis pipeline
Stars: ✭ 9 (-57.14%)
Mutual labels:  pipeline
Lambdacd
a library to define a continuous delivery pipeline in code
Stars: ✭ 655 (+3019.05%)
Mutual labels:  pipeline
Pipeline
A cloud-native Pipeline resource.
Stars: ✭ 6,751 (+32047.62%)
Mutual labels:  pipeline
Cookiecutter
DEPRECIATED! Please use nf-core/tools instead
Stars: ✭ 18 (-14.29%)
Mutual labels:  pipeline
Pdpipe
Easy pipelines for pandas DataFrames.
Stars: ✭ 590 (+2709.52%)
Mutual labels:  pipeline
Gollum
An n:m message multiplexer written in Go
Stars: ✭ 883 (+4104.76%)
Mutual labels:  pipeline
Galaxy
Data intensive science for everyone.
Stars: ✭ 812 (+3766.67%)
Mutual labels:  pipeline
Brunch
🍴 Web applications made easy. Since 2011.
Stars: ✭ 6,801 (+32285.71%)
Mutual labels:  pipeline
Aws Boilerplate
Opinionated full stack web app's boilerplate, ready to be deployed to AWS platform.
Stars: ✭ 682 (+3147.62%)
Mutual labels:  pipeline
Toil
A scalable, efficient, cross-platform (Linux/macOS) and easy-to-use workflow engine in pure Python.
Stars: ✭ 733 (+3390.48%)
Mutual labels:  pipeline
Segmentation
Catalyst.Segmentation
Stars: ✭ 27 (+28.57%)
Mutual labels:  pipeline
Bk Sops
蓝鲸智云标准运维(SOPS)
Stars: ✭ 632 (+2909.52%)
Mutual labels:  pipeline
Ppipe
A simple and lightweight Rust library for making iterator pipelines concurrent
Stars: ✭ 13 (-38.1%)
Mutual labels:  pipeline
Go Streams
A lightweight stream processing library for Go
Stars: ✭ 615 (+2828.57%)
Mutual labels:  pipeline
Phila Airflow
Stars: ✭ 16 (-23.81%)
Mutual labels:  pipeline
Steppy Toolkit
Curated set of transformers that make your work with steppy faster and more effective 🔭
Stars: ✭ 21 (+0%)
Mutual labels:  pipeline
Param pipe
parameterized pipe in elixir: |n>
Stars: ✭ 14 (-33.33%)
Mutual labels:  pipeline
Vector
A reliable, high-performance tool for building observability data pipelines.
Stars: ✭ 8,736 (+41500%)
Mutual labels:  pipeline

SPeW: SeqPipeWrap

Pitt-NCBI Hackathon Team Amanda Poholek (Lead), Argus Sun, Zhou Fang (Ark), Natalie Rittenhouse, Rahil Sethi, Ramya Mallela

Introduction

Taking a sequencing pipeline from individual pieces on your workstation to a seamless pipeline that can be run by any user is currently a challenge. Here, we create a proof-of-principle simple RNA-seq pipeline using separate Bash shell scripts that are linked together using NextFlow as a pipeline management tool. This is then wrapped using Docker for distribution and seamless running on other workstations without dependency issues.

SPeW is a framework for taking a NextGen Seq pipeline (such as RNA-seq, ChIP-seq or ATAC-seq) in any language, and using NextFlow as a pipeline management system to create a flexible, user-friendly pipeline that can be shared in a container platform.

This project was part of the September 2017 Pitt-NCBI Hackathon in Pittsburgh PA

Dependencies

Docker https://www.docker.com/

samtools http://www.htslib.org/

FastQC https://www.bioinformatics.babraham.ac.uk/projects/fastqc/

cutadapt https://github.com/marcelm/cutadapt

tophat2 http://cole-trapnell-lab.github.io/projects/tophat/

bowtie2 http://bowtie-bio.sourceforge.net/manual.shtml

cufflinks https://github.com/cole-trapnell-lab/cufflinks

R https://www.r-project.org/

NextFlow https://www.nextflow.io/

Methods

ScreenShot

To create the proof-of-principle simple RNA-seq pipeline, we started with writing simple Bash shell scripts for each step required in the analysis. These individual steps were then combined together by integrating them into Nextflow. In order to allow for seamless running on any workstation, Docker was then used to wrap the Nextflow code. By wrapping into a container such as Docker, all dependencies required for each step are automatically on the users workstation. Docker has the ability to be used by Singularity, allowing the code to be utilized on a High Computing Cluster(HPC).

Using Nextflow to String together Bash Scripts

Nextflow is easily installed by using curl:

curl -fsSL get.nextflow.io | bash

After successfully installing Nextflow, all the bash scripts are moved into a bin folder and the execute permission is granted for all the files.

mkdir -p bin 
mv *.sh bin
chmod +x *.sh

With all the bash files in the same folder, the Nextflow script can now be written. First, the environment is set and the initial input files are set as parameters:

#!/usr/bin/env nextflow

/*
* input parameters for the pipeline
*/
params.in = "$baseDir/data/*.fastq.gz"

A file object is then created from the string parameter:

inFiles = file(params.in)

By creating the file object, the file can now be used as the input file for the first process. Such as:

process trimming{

input: 
file reads from inFiles
}

The output specifies a file ('trimmed_*') that is then put into a variable(trimmedFiles) to be used in the next process.

process trimming{

input: 
file reads from inFiles

output:
file 'trimmed_*' into trimmedFiles
}

The script is now ready to be called. These scripts include the ability to use conditional statements:

process trimming{

input: 
file reads from inFiles

output:
file 'trimmed_*' into trimmedFiles

script:
singleEnd = true
adapter1 = "ADATPER_FWD"
adapter2 = "ADATPER_REV"

if (singleEnd==true)
"""
trimming.sh -i reads -s -a1 adapter1 -a2 adapter2
"""
else
"""
trimming.sh -i reads -a1 adapter1 -a2 adapter2
"""
}

The output from the trimming step of RNA-seq can easily be accessed by the aligning step:

process trimming{

input: 
file trim from trimmedFiles

output:
file '*.bam' into trimmedFiles

script:
align.sh trim

Nextflow seamlessly goes from one process to another by creating a 'work' folder in which all the intermediate files are placed. This work folder also ensures that if/when you run the script multiple times, the results will not be overwritten.

Discussion Notes

Overview

This repo was created as part of an NCBI-hackathon @ Univerisity of Pittsburgh in September of 2017. During this 3 day event, several disucssions occured about the best option to accomplish our goal of creating a method to take a NextGen Seq pipeline into a flexible format that can be easily shared. This is a summary of those discussions, edited by those who were present.

Most Seq pipelines are inherently a string of pre-made programs linked together by the creator according to their personal preferences or prior experience, including comfort of specific programming languages, knowledge of what algorithms exist, and word of mouth about what is "best" option to use. This typically limits sharing the pipeline to another user who may want to make small modifications, or requires installing new software in order to run. In addition, many users may choose to run their pipeline on a local machine, cloud computing or a HPC. Thus, putting a pipeline into a format or framework that manages the pipeline and allows for easy modifications for future users as well as ability to share, would be of tremendous use to many beginner or intermediate bioinformatics users.

Test case

As a starting point, we generated a basic RNA-seq pipeline that was made up of individual modules in a bash shell script. These modules are as follows:

  1. FastQC
  2. Trimming
  3. Alignment
  4. Annotation
  5. Differential Gene Expression (DGE)

Inputs and outputs were defined in Nextflow to link processess together. All outputs were placed in a final directory that had final products as well as intermediate files.

Workflow Management Strategy Discussion with a Group of ~25 Computational Biologists and Data Scientists

Next, we needed to determine a method to link these modules together where inputs and outputs are managed for you, and that a future user could add or change a module. Our ideal scenario includes an option where the pipeline can make decisions for you. For example, given output X, th next module used where output X becomes input X could be either module Y or module Z. In addition, we considered how to package the pipeline. This included the option of either wrapping each module in a container (ie docker) that then are linked together within a larger container, or making one whole pipeline and putting it into one container.

The majority of the discussion was around how to link the modules together. Several pipeline management systems were discussed, including Nextflow, snakemake, CWL (common workflow language) and Jupyter notebooks. This conversation included the entire Hackathon group (5 teams) of roughly 25 people total. Ultimately we selected Nextflow, based on the fact that we think Nextflow has the best options for our needs. There was not universal agreement about this, and we believe this will a trial and error process. One downside was no one in the room had any experience with Nextflow, thus we were testing to see if this was indeed a good option.

CWL was widely dismissed by pretty much all members present, as being too labor intensive to use. A few people with CWL experience relayed how difficult and frustrating it was to use, and the time it took to learn considered not worth the effort.

Snakemake was dismissed as being less flexible than Nextflow. Many users thought that it is mostly Python oriented, although others confirmed that is not the case.

Nextflow was chosen because it can use any language, manages inputs and outputs and is meant to be easily wrapped.

A large part of the discussion included Jupyter notebooks as an alternative to Nextflow. This was considered to be a good in-between for intermediate-level bioinformaticians who want to crack the containers and customize them for particular use cases. Many raised issues of needing to install Jupyter, however no installation should be necessary if Jupyter is "headless" in a container, and Jupyter would only be used by those cracking the containers. However, the we feel it is important to be able to encompass all languages, and therefore this option may have inherent limitations, but perhaps be attractive for others in the future.

Based on these discussions, we chose Nextflow and here report how useful and flexible it ended up being. We hope this may help other users make more informed choices to manage their pipelines and wrap for distribution.

Futher Directions

We hope to add additional modules as well as modify current modules to the pipeline, and add in the ability to enter the pipeline at any point.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].