Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enhancing documentation for viral-pipelines #461

Merged
merged 51 commits into from
Jul 19, 2023
Merged
Show file tree
Hide file tree
Changes from 25 commits
Commits
Show all changes
51 commits
Select commit Hold shift + click to select a range
0921e34
Adding traning workflow wdl and changing workflows.rst to add demux_d…
golu099 Mar 2, 2023
b3307de
modifying 16S-trian_classifier.wdl
golu099 Mar 2, 2023
166749d
addressing same call issue
golu099 Mar 2, 2023
6df4826
fixing bug issue
golu099 Mar 2, 2023
99102f1
16S train classifier bugs
golu099 Mar 2, 2023
20ece88
tax ref bug
golu099 Mar 2, 2023
92c76fc
tax debug
golu099 Mar 2, 2023
5d53f4b
Description meta
golu099 Mar 23, 2023
355555a
Fixes with task_assembly
golu099 Mar 30, 2023
6d15014
bug fixes in assembly task
golu099 Mar 30, 2023
492e867
bug fixes in assembly task
golu099 Mar 30, 2023
6f8eae2
bug fixes in assembly task
golu099 Mar 30, 2023
072763d
Addressing line 21
golu099 Mar 30, 2023
c59384f
addressing bug fixes for task assembly
golu099 Mar 30, 2023
0e22527
addressing bug fixes for task assembly
golu099 Mar 30, 2023
b0eca2c
addressing line 71
golu099 Mar 30, 2023
661b286
Adding documentation to demux_deplete
golu099 Apr 5, 2023
f63c457
fixing comma issue in task_report
golu099 Apr 5, 2023
5a3ed7a
Addressing commas
golu099 Apr 6, 2023
c2fe9f4
Addressing commas
golu099 Apr 6, 2023
a801357
Removing novalaign license
golu099 Apr 6, 2023
b5a19fe
Merge branch 'master' into fnegrete_test
dpark01 Apr 13, 2023
d3dff3c
Update docs/workflows.rst
golu099 Apr 13, 2023
2daf478
Removing .vscode files from git
golu099 Apr 20, 2023
8ccf287
changing dockstore.yml
golu099 Apr 26, 2023
cfbd797
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
e3c34b0
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
a44066d
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
8825191
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
de88be5
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
2c25808
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
ec81e59
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
b9e5cdc
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 Apr 26, 2023
b699894
Update pipes/WDL/workflows/demux_deplete.wdl
golu099 Apr 26, 2023
2ebbebb
chaning the memory storage
golu099 Apr 27, 2023
07df9d4
chaning memory storage for amplicon wdl
golu099 Apr 27, 2023
232019d
Merge branch 'fnegrete_test' of https://github.com/broadinstitute/vir…
golu099 Apr 27, 2023
1a6a255
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 May 2, 2023
ef2f409
Update pipes/WDL/tasks/tasks_16S_amplicon.wdl
golu099 May 2, 2023
b652ad0
Update pipes/WDL/tasks/tasks_16S_amplicon.wdl
golu099 May 2, 2023
9e5cfcc
Update pipes/WDL/tasks/tasks_16S_amplicon.wdl
golu099 May 2, 2023
9366952
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 May 2, 2023
f1321fd
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 May 2, 2023
97e40d2
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 May 2, 2023
e88d81d
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 May 2, 2023
fe29c46
Update pipes/WDL/tasks/tasks_assembly.wdl
golu099 May 2, 2023
1482761
Update pipes/WDL/tasks/tasks_16S_amplicon.wdl
golu099 May 2, 2023
478b7ad
Update pipes/WDL/tasks/tasks_16S_amplicon.wdl
golu099 May 2, 2023
a8281e6
Merge branch 'master' into fnegrete_test
tomkinsc May 11, 2023
36f7854
Merge branch 'master' into fnegrete_test
tomkinsc May 19, 2023
0eb3b71
Merge branch 'master' into fnegrete_test
tomkinsc Jul 19, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions .dockstore.yml
Original file line number Diff line number Diff line change
Expand Up @@ -349,13 +349,13 @@ workflows:
primaryDescriptorPath: /pipes/WDL/workflows/trimal.wdl
testParameterFiles:
- empty.json
- name: amplicon16S_analysis
- name: classify_qiime2_multi
subclass: WDL
primaryDescriptorPath: /pipes/WDL/workflows/amplicon16S_analysis.wdl
primaryDescriptorPath: /pipes/WDL/workflows/classify_qiime2_multi.wdl
testParameterFiles:
- empty.json
- name: qiime_import_bam
subclass: WDL
primaryDescriptorPath: /pipes/WDL/workflows/qiime_import_bam.wdl
primaryDescriptorPath: /pipes/WDL/workflows/bam_to_qiime.wdl
testParameterFiles:
- empty.json
2 changes: 2 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -77,3 +77,5 @@ tools/conda-tools/

# miniWDL
default.profraw

.vscode/
4 changes: 2 additions & 2 deletions docs/workflows.rst
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ WDL Workflows

Documentation for each workflow is provided here. Although there are many workflows that serve different functions, some of the primary workflows we use most often include:

- :doc:`demux_plus` (on every sequencing run)
- :doc:`classify_multi` (metagenomics & host depletion)
- :doc:`demux_deplete` (on every sequencing run)
- :doc:`classify_single` (metagenomics & host depletion)
- :doc:`assemble_denovo` (for most viruses)
- :doc:`assemble_refbased` (for less diverse viruses, such as those from single point source human outbreaks)
- :doc:`augur_from_assemblies` (for nextstrain-based visualization of phylogeny)
Expand Down
144 changes: 137 additions & 7 deletions pipes/WDL/tasks/tasks_16S_amplicon.wdl
Original file line number Diff line number Diff line change
Expand Up @@ -6,14 +6,24 @@ task qiime_import_from_bam {
}
input {
Array[File] reads_bam
Int memory_mb = 7000
Int memory_mb = 7000
golu099 marked this conversation as resolved.
Show resolved Hide resolved
Int cpu = 5
Int disk_size_gb = ceil(2*20) + 5
String docker = "quay.io/broadinstitute/qiime2"
}
parameter_meta {
reads_bam: {description: "Unaligned reads in BAM format, one sample per BAM file."}
reads_qza: {description: "All unaligned reads in a single QZA (QIIME) file"}
reads_bam: {
description: "Unaligned reads in BAM format, one sample per BAM file.",
category: "required"
}
reads_qza: {
description: "All unaligned reads in a single QZA (QIIME) file.",
category: "other"
}
samplename_master_sheet: {
description: "File contains all samples names.",
category: "other"
}
}

command <<<
Expand Down Expand Up @@ -69,17 +79,45 @@ task trim_reads {

input {
File reads_qza
#Boolean not_default = false
String forward_adapter = "CTGCTGCCTCCCGTAGGAGT"
String reverse_adapter = "AGAGTTTGATCCTGGCTCAG"
Int min_length = 1
Int min_length = 1
golu099 marked this conversation as resolved.
Show resolved Hide resolved
Boolean keep_untrimmed_reads = false
Int memory_mb = 2000
Int cpu = 4
Int disk_size_gb = ceil(2*size(reads_qza, "GiB")) + 5
String docker = "quay.io/broadinstitute/qiime2"
}

parameter_meta {
reads_qza: {
description: "All unaligned reads in a single QZA (QIIME) file.",
cateogry: "required"
}
forward_adapter: {
description: "Forward amplicon primer sequence.",
category: "advanced"
}
reverse_adapter: {
description: "Reverse amplicon primer sequence.",
cateogry: "advanced"
}
min_length: {
description: "Minimum length of the read, cutadapt will discard anything that is shorter than n bp AFTER trimming.Set to default.",
Copy link
Member

@tomkinsc tomkinsc Jul 18, 2023

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Whose default? qiime's? cutadapt's? Maybe specify the source of the default and/or its value here. Also maybe specify that it would be reads that would be discarded: "discard anything that is" -> "discard reads that are". If such reads are discarded, does that extend to both reads of a pair for paired-end sequencing, or could it toss one read and keep its mate?

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

After this small change, let's get this merged in.

category: "other"
}
keep_untrimmed_reads: {
description: "Allows you to choose whether or not to discard untrimmed reads.",
category: "advanced"
}
trimmed_reads_qza: {
description: "Trimmed reads data file.",
category: "advanced"
}
trimmed_visualization: {
description: "A diagram that compares your demuxed reads before and after cutting (i.e. length of reads, how many reads were retained).",
category: "advanced"
}
}
command <<<
set -ex -o pipefail
qiime cutadapt trim-paired \
Expand Down Expand Up @@ -124,7 +162,20 @@ task join_paired_ends {
Int disk_size_gb = ceil(2*size(trimmed_reads_qza, "GiB")) + 50
String docker = "quay.io/broadinstitute/qiime2"
}

parameter_meta{
trimmed_reads_qza: {
description:"Trimmed reads data file.",
category: "required"
}
joined_end_reads_qza:{
description: "Merge paired read file.",
category: "other"
}
joined_end_visualization: {
description: "This summary is especially useful for assessing the length of linked reads and the quality scores at each sequence base position. ",
category: "other"
}
}
command <<<
set -ex -o pipefail
qiime vsearch join-pairs \
Expand Down Expand Up @@ -160,6 +211,32 @@ task deblur {
Int cpu = 1
Int disk_size_gb = ceil(2*size(joined_end_reads_qza, "GiB")) + 5
String docker = "quay.io/broadinstitute/qiime2"
}
parameter_meta {
joined_end_reads_qza: {
description: "Merge paired read file.",
category: "required"
}
trim_length_var: {
description: "Length that all seqeuences will be trimmed, and discard any sequences that are not at least this long.",
category: "advanced"
}
representative_seqs_qza: {
description: "Generate a list of the representative sequences. May be useful to the user if they want to blast these sequences or check for correct trimming.",
category: "other"
}
representative_table_qza: {
description: "Generate a table of the representaitve sequences.",
category: "other"
}
feature_table: {
description: "A table that represent the number of of features per sample, the number of samples a given feature is found in.",
category: "other"
}
visualize_stats:{
description: "Generate visualization of deblur stats.",
category: "other"
}
}
command <<<
set -ex -o pipefail
Expand Down Expand Up @@ -213,6 +290,37 @@ task train_classifier {
Int disk_size_gb = ceil(2*size(otu_ref, "GiB")) + 5
String docker = "quay.io/broadinstitute/qiime2"
}
parameter_meta{
otu_ref: {
description: "Operational taxonomic units (OTUs) sequences imported as FASTA file.",
category:"required"
}
taxanomy_ref: {
description: "Reference taxonomy file.",
category: "required"
}
forward_adapter: {
description: "The forward primer sequence for the amplicon target.",
category: "advanced"
}
reverse_adapter: {
description: "The reverse primer sequence for the amplicon target.",
category:"advanced"
}
min_length: {
description: "Minimum length of amplicon sequences.",
category: "advanced"
}
max_length: {
description: "Maximum length of amplicon sequences.",
category:"advanced"
}
trained_classifier: {
description: "Trained taxonomic classifier on target amplicon sequences.",
category: "other"
}
}

command <<<
set -ex -o pipefail
CONDA_ENV_NAME=$(conda info --envs -q | awk -F" " '/qiime.*/{ print $1 }')
Expand Down Expand Up @@ -266,6 +374,28 @@ task tax_analysis {
Int disk_size_gb = 375
String docker = "quay.io/broadinstitute/qiime2"
}
parameter_meta{
trained_classifier: {
description: "Trained taxonomic classifier on target amplicon sequences.",
category: "required"
}
representative_seqs_qza: {
description: "List of representative sequences.",
category:"required"
}
representative_table_qza: {
description: "Table of representative sequences.",
category:"other"
}
rep_seq_list: {
description: "Generate list of representative sequences.",
category:"other"
}
tax_classification_graph: {
description: "Create a bar graph of your taxonomic classification.",
category:"other"
}
}
command <<<
set -ex -o pipefail
qiime feature-classifier classify-sklearn \
Expand Down
Loading