-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathnextflow.config
95 lines (81 loc) · 2.91 KB
/
nextflow.config
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
manifest {
name = 'nf-eggnog'
author = 'Sunit Jain'
homePage = 'https://github.com/FischbachLab/nf-eggnog'
description = 'Run EggNOG-mapper on a set of protein sequences.'
mainScript = 'main.nf'
defaultBranch = 'main'
version = '1.0.0'
}
// TODO: Add the ability to switch profiles from the command line. Make the AWS Batch profile the default.
// Until I get around doing this, Uncomment the following block to run the pipeline locally. and comment out the includeConfig line.
// Uncomment the following block for local testing
// profiles {
// local {
// process.executor = 'local'
// docker.enabled = true
// }
// awsbatch {
// includeConfig "conf/aws_batch.config"
// }
// }
// Comment the following line for local testing
includeConfig "$projectDir/conf/aws_batch.config"
params {
/*
* Defines the pipeline inputs parameters (giving a default value for each for them)
* Each of the following parameters can be specified as command line options
*/
outdir = "s3://genomics-workflow-core/Results/EggNOG"
project = "00_Test"
prefix = "20241212"
seedfile = null
workflow = null
// docker containers
eggnog_container = "458432034220.dkr.ecr.us-west-2.amazonaws.com/eggnog:2.1.12"
db_tarball = "s3://genomics-workflow-core/Results/EggNOG/00_Database/v5.0.2.tar.gz"
db_path = "v5.0.2"
}
process {
maxRetries = 3
errorStrategy = { task.exitStatus in [104,134,137,138,139,140,143,199] ? 'retry' : 'finish' }
withLabel: 'large_compute' {
queue = 'highmem-ec2-maf-pipelines'
time = 8.hour
cpus = {30 * task.attempt}
memory = {110.GB * task.attempt}
container = params.eggnog_container
}
withLabel: 'small_compute' {
queue = 'default-maf-pipelines'
time = 4.hour
cpus = {7 * task.attempt}
memory = {28.GB * task.attempt}
container = params.eggnog_container
}
}
// This seems to be quite unreliable when running on AWS Batch
// some jobs run just fine, where as other jobs fail with the following error:
// "Command 'ps' required by nextflow to collect task metrics cannot be found"
// Could be an AWS Batch issue where the container is not refreshed, maybe? not sure.
// removing this section for now.
// trace {
// enabled = false
// overwrite = true
// file = "${params.outdir}/${params.project}/${params.prefix}/${params.workflow}/reports/trace-${params.prefix}.tsv"
// }
// timeline {
// enabled = true
// overwrite = true
// file = "${params.outdir}/${params.project}/${params.prefix}/${params.workflow}/reports/timeline-${params.prefix}.html"
// }
// report {
// enabled = true
// overwrite = true
// file = "${params.outdir}/${params.project}/${params.prefix}/${params.workflow}/reports/execution-report-${params.prefix}.html"
// }
// dag {
// enabled = true
// overwrite = true
// file = "${params.outdir}/${params.project}/${params.prefix}/${params.workflow}/reports/dag-${params.prefix}.html"
// }