nf-core/multiplesequencealign
A pipeline to run and systematically evaluate Multiple Sequence Alignment (MSA) methods.
Introduction
This document describes the output produced by the pipeline.
The directories listed below will be created in the results directory after the pipeline has finished. All paths are relative to the top-level results directory.
Pipeline overview
The pipeline is built using Nextflow and processes data using the following steps:
-
Input files summary: (Optional) computation of summary statistics on the input fasta file, such as the average sequence similarity across the input sequences, their length, etc. Skip by
--skip_stats
as a parameter. -
Guide Tree: (Optional) Renders a guide tree. This is meant to try different combinations of guide trees and assembly tools, if you are interested in a standard alignment procedure you can ignore this step.
-
Align: aligns the sequences.
-
Evaluate: (Optional) The obtained alignments are evaluated with different metrics: Sum Of Pairs (SoP), Total Column score (TC), iRMSD, Total Consistency Score (TCS), etc. Skip by passing
--skip_eval
as a parameter. -
Report: Reports about the collected information of the runs are reported in a shiny app and a summary table in multiqc. Skip by passing
--skip_shiny
and--skip_multiqc
.
Input files summary
Statistics about the input files are collected and summarized into a final csv file.
Output files
summary/stats/
complete_summary_stats.csv
: csv file containing the summary for all the statistics computed on the input file.sequences/
seqstats/*_seqstats.csv
: file containing the sequence input length for each sequence in the family defined by the file name. If--calc_seq_stats
is specified.perc_sim/*_txt
: file containing the pairwise sequence similarity for all input sequences. If--calc_sim
is specified.
structures/
-plddt/*_full_plddt.csv
: file containing the plddt of the structures for each sequence in the input file. If--extract_plddt
is specified.
Trees
If you explicitly specifified (via the toolsheet) to compute guidetrees to be used by the MSA tool, those are stored in the trees directory.
Output files
trees/
*/*.dnd
: guide tree files.
Alignment
All MSA computed are stored in the alignment directory.
Output files
alignment/
*/*.fa
: each subdirectory is named after the sample id. It contains all the alignments computed on it. The filename contains all the informations of the input file used and the tool. The file naming convention is: {Inputfile}{Tree}_args-{Tree_args}_{MSA}_args-{MSA_args}.aln
Evaluation
Files with the summary of the computed evaluation statistics are stored in the evaluation directory.
Output files
evaluation/
tcoffee_irmsd/
: directory containing the files with the complete iRMSD files. If--calc_irmsd
is specified.tcoffee_tcs/
: directory containing the files with the complete TCS files. If--calc_tcs
is specified.complete_summary_eval.csv
: csv file containing the summary of all evaluation metrics for each input file.
Shiny App
A Shiny app is created to explore interactively your results. It can be found in the shiny_app folder.
A shiny app is prepared to visualize the summary statistics and evaluation of the produced alignments (skip with --skip_shiny
).
To run the Shiny app use the following commands from the results directory:
cd shiny_app
./run.sh
Output files
shiny_app/
run.sh
: executable to start the shiny app.*.py*
: shiny app files.*.csv
: csv file used by shiny app.trace.txt
: trace file used by shiny app.
Be aware that you have to have shiny installed to access this feature.
MultiQC
You can find the MultiQC report in the multiqc
folder.
Output files
multiqc/
multiqc_report.html
: a standalone HTML file that can be viewed in your web browser.multiqc_data/
: directory containing parsed statistics from the different tools used in the pipeline.multiqc_plots/
: directory containing static images from the report in various formats.
MultiQC is a visualization tool that generates a single HTML report summarising all samples in your project. A table with all the collected statistics and evaluted metrics is reported as well as all the versions of the tools used for the computation.
Pipeline information
Extra information about the pipeline execution are stored in the pipeline_info folder.
Output files
pipeline_info/
- Reports generated by Nextflow:
execution_report.html
,execution_timeline.html
,execution_trace.txt
andpipeline_dag.dot
/pipeline_dag.svg
. - Reports generated by the pipeline:
pipeline_report.html
,pipeline_report.txt
andsoftware_versions.yml
. Thepipeline_report*
files will only be present if the--email
/--email_on_fail
parameter’s are used when running the pipeline. - Reformatted samplesheet files used as input to the pipeline:
samplesheet.valid.csv
. - Parameters used by the pipeline run:
params.json
.
- Reports generated by Nextflow:
Nextflow provides excellent functionality for generating various reports relevant to the running and execution of the pipeline. This will allow you to troubleshoot errors with the running of the pipeline, and also provide you with other information such as launch commands, run times and resource usage.