-
Laurent Modolo authoredLaurent Modolo authored
- Initialize your own project
- Forking
- Project organisation
- docker_modules
- psmn_modules
- nf_modules
- Nextflow pipeline
- Processes
- Channels
- Run your pipeline locally
- Getting your results
- Fasta everywhere
- Build your own RNASeq pipeline
- Create your Docker containers
- Cutadapt
- cutadapt.config
- UrQt
- BEDtools
- Kallisto
- Additional nextflow option
- Run your RNASeq pipeline on the PSMN
- Set your environment
- Run nextflow
title: "TP for experimental biologists"
author: Laurent Modolo [laurent.modolo@ens-lyon.fr](mailto:laurent.modolo@ens-lyon.fr)
date: 6 Jun 2018
output:
pdf_document:
toc: true
toc_depth: 3
number_sections: true
highlight: tango
latex_engine: xelatex
The Goal of this practical is to learn how to build your own pipeline with nextflow and using the tools already wrapped. For this we are going to build a small RNASeq analysis pipeline that should run the following steps:
- remove Illumina adaptors
- trim reads by quality
- build the index of a reference genome
- estimate the amount of RNA fragments mapping to the transcripts of this genome
Initialize your own project
You are going to build a pipeline for you or your team. So the first step is to create your own project.
Forking
Instead of reinventing the wheel, you can use the pipelines/nextflow as a template. To easily do so, go to the pipelines/nextflow repository and click on the fork button.
In git, the action of forking means that you are going to make your own private copy of a repository. You can then write modifications in your project, and if they are of interest for the source repository (here pipelines/nextflow) create a merge request. Merge requests are sent to the source repository to ask the maintainers to integrate modifications.
Project organisation
This project (and yours) follows the guide of good practices for the LBMC
You are now on the main page of your fork of the pipelines/nextflow. You can explore this project, all the code in it is under the CeCILL licence (in the LICENCE file).
The README.md file contains instructions to run your pipeline and test its installation.
The CONTRIBUTING.md file contains guidelines if you want to contribute to the pipelines/nextflow (making a merge request for example).
The data folder will be the place where you store the raw data for your analysis. The results folder will be the place where you store the results of your analysis.
The content of
data
andresults
folders should never be saved on git.
The doc folder contains the documentation of this practical course.
And most interestingly for you, the src contains code to wrap tools. This folder contains two subdirectories. A docker_modules
, a nf_modules
and a psmn_modules
folder.
docker_modules
The src/docker_modules
contains the code to wrap tools in Docker. Docker is a framework that allows you to execute software within containers. The docker_modules
contains directory corresponding to tools and subdirectories corresponding to their version.
ls -l src/docker_modules/
rwxr-xr-x 3 laurent _lpoperator 96 May 25 15:42 bedtools/
drwxr-xr-x 4 laurent _lpoperator 128 Jun 5 16:14 bowtie2/
drwxr-xr-x 3 laurent _lpoperator 96 May 25 15:42 fastqc/
drwxr-xr-x 4 laurent _lpoperator 128 Jun 5 16:14 htseq/
To each tools/version
corresponds two files:
ls -l src/docker_modules/bowtie2/2.3.4.1/
-rw-r--r-- 1 laurent _lpoperator 283 Jun 5 15:07 Dockerfile
-rwxr-xr-x 1 laurent _lpoperator 79 Jun 5 16:18 docker_init.sh*
The Dockerfile
is the Docker recipe to create a container containing Bowtie2
in its 2.3.4.1
version. And the docker_init.sh
file is a small script to create the container from this recipe.
By running this script you will be able to easily install tools in different versions on your personal computer and use it in your pipeline. Some of the advantages are:
- Whatever the computer, the installation and the results will be the same
- You can keep container for old version of tools and run it on new systems (science = reproducibility)
- You don’t have to bother with tedious installation procedures, somebody else already did the job and wrote a
Dockerfile
. - You can easily keep containers for different version of the same tools.
psmn_modules
The src/psmn_modules
folder is not really there. It’s a submodule of the project PSMN/modules. To populate it locally you can use the following command:
git submodule init
Like the src/docker_modules
the PSMN/modules project describe recipes to install tools and use them. The main difference is that you cannot use Docker on the PSMN. Instead you have to use another framework Environment Module which allows you to load modules for specific tools and version.
The README.md file of the PSMN/modules repository contains all the instruction to be able to load the modules maintained by the LBMC and present in the PSMN/modules repository.
nf_modules
The src/nf_modules
folder contains templates of nextflow wrappers for the tools available in Docker and psmn. The details of the nextflow wrapper will be presented in the next section. Alongside the .nf
and .config
files, there is a tests.sh
script to run test on the tool.
Nextflow pipeline
A pipeline is a succession of process. Each process has data input(s) and optional data output(s). Data flows are modeled as channels.
Processes
Here is an example of process:
process sample_fasta {
input:
file fasta from fasta_file
output:
file "sample.fasta" into fasta_sample
script:
"""
head ${fasta} > sample.fasta
"""
}
We have the process sample_fasta
that takes a fasta_file
channel as input and as output a fasta_sample
channel. The process itself is defined in the script:
block and within """
.
input:
file fasta from fasta_file
When we zoom on the input:
block we see that we define a variable fasta
of type file
from the fasta_file
channel. This mean that groovy is going to write a file named as the content of the variable fasta
in the root of the folder where script:
is executed.
output:
file "sample.fasta" into fasta_sample
At the end of the script, a file named sample.fasta
is found in the root the folder where script:
is executed and send into the channel fasta_sample
.
Using the WebIDE of Gitlab, create a file src/fasta_sampler.nf
with this process and commit it to your repository.
Channels
Why bother with channels? In the above example, the advantages of channels are not really clear. We could have just given the fasta
file to the process. But what if we have many fasta files to process? What if we have sub processes to run on each of the sampled fasta files? Nextflow can easily deal with these problems with the help of channels.
Channels are streams of items that are emitted by a source and consumed by a process. A process with a channel as input will be run on every item send through the channel.
Channel
.fromPath( "data/tiny_dataset/fasta/*.fasta" )
.set { fasta_file }
Here we defined the channel fasta_file
that is going to send every fasta file from the folder data/tiny_dataset/fasta/
into the process that take it as input.
Add the definition of the channel to the src/fasta_sampler.nf
file and commit it to your repository.
Run your pipeline locally
After writing this first pipeline, you may want to test it. To do that, first clone your repository. To easily do that set the visibility level to public in the settings/General/Permissions page of your project.
You can then run the following commands to download your project on your computer:
If you are on a PSMN computer:
pip install cutadapt=1.14
PATH="/scratch/lmodolo/:$PATH"
git config --global http.sslVerify false
and then :
Don't forget to replace https://gitlab.biologie.ens-lyon.fr/ by gitlab_lbmc if you are using your own computer
git clone https://gitlab.biologie.ens-lyon.fr/<usr_name>/nextflow.git
cd nextflow
src/install_nextflow.sh
We also need data to run our pipeline:
cd data
git clone https://gitlab.biologie.ens-lyon.fr/LBMC/tiny_dataset.git
cd ..
We can run our pipeline with the following command:
./nextflow src/fasta_sampler.nf
Getting your results
Our pipeline seems to work but we don’t know where is the sample.fasta
. To get results out of a process, we need to tell nextflow to write it somewhere (we may don’t need to get every intermediate file in our results).
To do that we need to add the following line before the input:
section:
publishDir "results/sampling/", mode: 'copy'
Every file described in the output:
section will be copied from nextflow to the folder results/sampling/
.
Add this to your src/fasta_sampler.nf
file with the WebIDE and commit to your repository.
Pull your modifications locally with the command:
git pull origin master
You can run your pipeline again and check the content of the folder results/sampling
.
Fasta everywhere
We ran our pipeline on one fasta file. How would nextflow handle 100 of them? To test that we need to duplicate the tiny_v2.fasta
file:
for i in {1..100}
do
cp data/tiny_dataset/fasta/tiny_v2.fasta data/tiny_dataset/fasta/tiny_v2_${i}.fasta
done
You can run your pipeline again and check the content of the folder results/sampling
.
Every fasta_sampler
process write a sample.fasta
file. We need to make the name of the output file dependent of the name of the input file.
output:
file "*_sample.fasta" into fasta_sample
script:
"""
head ${fasta} > ${fasta.baseName}_sample.fasta
"""
Add this to your src/fasta_sampler.nf
file with the WebIDE and commit it to your repository before pulling your modifications locally.
You can run your pipeline again and check the content of the folder results/sampling
.
Build your own RNASeq pipeline
In this section you are going to build your own pipeline for RNASeq analysis from the code available in the src/nf_modules
folder.
Create your Docker containers
For this practical, we are going to need the following tools:
- For Illumina adaptor removal: cutadapt
- For reads trimming by quality: UrQt
- For mapping and quantifying reads: BEDtools and Kallisto
To initialize these tools, follow the Installing section of the README.md file.
If you are using a CBP computer don’t forget to clean up your docker containers at the end of the practical with the following commands:
docker rm $(docker stop $(docker ps -aq))
docker rmi $(docker images -qf "dangling=true")
Cutadapt
The first step of the pipeline is to remove any Illumina adaptors left in your read files.
Open the WebIDE and create a src/RNASeq.nf
file. Browse for src/nf_modules/cutadapt/adaptor_removal_paired.nf, this file contains examples for cutadapt. We are interested in the Illumina adaptor removal, for paired-end data section of the code. Copy this code in your pipeline and commit it.
Compared to before, we have few new lines:
params.fastq = "$baseDir/data/fastq/*_{1,2}.fastq"
We declare a variable that contains the path of the fastq file to look for. The advantage of using params.fastq
is that the option --fastq
is now a parameter of your pipeline.
Thus, you can call your pipeline with the --fastq
option:
./nextflow src/RNASeq.nf --fastq "data/tiny_dataset/fastq/*_R{1,2}.fastq"
log.info "fastq files: ${params.fastq}"
This line simply displays the value of the variable
Channel
.fromFilePairs( params.fastq )
As we are working with paired-end RNASeq data, we tell nextflow to send pairs of fastq in the fastq_file
channel.
cutadapt.config
For the fastq_sampler.nf
pipeline we used the command head
present in most base UNIX systems. Here we want to use cutadapt
which is not. Therefore, we have three main options:
- install cutadapt locally so nextflow can use it
- launch the process in a Docker container that has cutadapt installed
- launch the process with psmn while loading the correct module to have cutadapt available
We are not going to use the first option which requires no configuration for nextflow but tedious tools installations. Instead, we are going to use existing wrappers and tell nextflow about it. This is what the src/nf_modules/cutadapt/adaptor_removal_paired.config is used for.
Copy the content of this config file to an src/RNASeq.config
file. This file is structured in process blocks. Here we are only interested in configuring adaptor_removal
process not trimming
process. So you can remove the trimming
block and commit it.
You can test your pipeline with the following command:
./nextflow src/RNASeq.nf -c src/RNASeq.config -profile docker --fastq "data/tiny_dataset/fastq/*_R{1,2}.fastq"
UrQt
The second step of the pipeline is to trim reads by quality.
Browse for src/nf_modules/urqt/trimming_paired.nf, this file contains examples for UrQt. We are interested in the for paired-end data section of the code. Copy the process section code in your pipeline and commit it.
This code won’t work if you try to run it: the fastq_file
channel is already consumed by the adaptor_removal
process. In nextflow once a channel is used by a process, it ceases to exist. Moreover, we don’t want to trim the input fastq, we want to trim the fastq that comes from the adaptor_removal
process.
Therefore, you need to change the line:
set pair_id, file(reads) from fastq_files
In the trimming
process to:
set pair_id, file(reads) from fastq_files_cut
The two processes are now connected by the channel fastq_files_cut
.
Add the content of the src/nf_modules/urqt/trimming_paired.config file to your src/RNASeq.config
file and commit it.
You can test your pipeline.
BEDtools
Kallisto need the sequences of the transcripts that need to be quantified. We are going to extract these sequences from the reference data/tiny_dataset/fasta/tiny_v2.fasta
with the bed
annotation data/tiny_dataset/annot/tiny.bed
.
You can copy to your src/RNASeq.nf
file the content of src/nf_modules/bedtools/fasta_from_bed.nf and to your src/RNASeq.config
file the content of src/nf_modules/bedtools/fasta_from_bed.config.
Commit your work and test your pipeline with the following command:
./nextflow src/RNASeq.nf -c src/RNASeq.config -profile docker --fastq "data/tiny_dataset/fastq/*_R{1,2}.fastq" --fasta "data/tiny_dataset/fasta/tiny_v2.fasta" --bed "data/tiny_dataset/annot/tiny.bed"
Kallisto
Kallisto run in two steps: the indexation of the reference and the quantification on this index.
You can copy to your src/RNASeq.nf
file the content of the files src/nf_modules/kallisto/indexing.nf and src/nf_modules/kallisto/mapping_paired.nf. You can add to your file src/RNASeq.config
file the content of the files src/nf_modules/kallisto/indexing.config and src/nf_modules/kallisto/mapping_paired.config.
We are going to work with paired-end so only copy the relevant processes. The index_fasta
process needs to take as input the output of your fasta_from_bed
process. The fastq
input of your mapping_fastq
process needs to take as input the output of your index_fasta
process and the trimming
process.
Commit your work and test your pipeline. You now have a RNASeq analysis pipeline that can run locally with Docker!
Additional nextflow option
With nextflow you can restart the computation of a pipeline and get a trace of the process with the following options:
-resume -with-dag results/RNASeq_dag.pdf -with-timeline results/RNASeq_timeline
Run your RNASeq pipeline on the PSMN
First you need to connect to the PSMN:
login@allo-psmn
Then once connected to allo-psmn
, you can connect to e5-2667v4comp1
:
login@e5-2667v4comp1
Set your environment
Make the LBMC modules available to you:
ln -s /Xnfs/lbmcdb/common/modules/modulefiles ~/privatemodules
echo "module use ~/privatemodules" >> .bashrc
Create and go to your scratch
folder:
mkdir -p /scratch/<login>
cd /scratch/<login>
echo "module use ~/privatemodules" >> .bashrc
Then you need to clone your pipeline and get the data:
git config --global http.sslVerify false
git clone https://gitlab.biologie.ens-lyon.fr/<usr_name>/nextflow.git
cd nextflow/data
git clone https://gitlab.biologie.ens-lyon.fr/LBMC/tiny_dataset.git
cd ..
Run nextflow
As we don’t want nextflow to be killed in case of disconnection, we start by launching tmux
. In case of deconnection, you can restore your session with the command tmux a
.
tmux
module load nextflow/0.28.2
nextflow src/RNASeq.nf -c src/RNASeq.config -profile psmn --fastq "data/tiny_dataset/fastq/*_R{1,2}.fastq" --fasta "data/tiny_dataset/fasta/tiny_v2.fasta" --bed "data/tiny_dataset/annot/tiny.bed" -w /scratch/<login>
To use the scratch for nextflow computations add the option :
-w /scratch/<login>
You just ran your pipeline on the PSMN!