-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathnextflow.config
161 lines (127 loc) · 5.59 KB
/
nextflow.config
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
params {
msconvert.do_demultiplex = false; // whether or not to demultiplex with msconvert
msconvert.do_simasspectra = true; // whether or not to do simAsSpectra with msconvert
msconvert.store_mzml = false;
pdc_base_url = 'https://proteomic.datacommons.cancer.gov/graphql'
pdc_client_args = ''
pdc_study_id = null
n_raw_files = null
pdc_annotations_csv = null
pdc_metadata_tsv = null
ms_data_dir = null // If null, all the raw files for the study through the PDC API.
// general workflow params, can be changed
result_dir = 'results/nf-pdc-dia'
report_dir = 'reports/nf-pdc-dia'
search_engine = 'encyclopedia'
// default parameters for Encyclopedia searches, can be overridden by users
encyclopedia.params = '-enableAdvancedOptions -v2scoring'
encyclopedia.version = '3.0.0-SNAPSHOT'
// fasta = 'https://panoramaweb.org/_webdav/MacCoss/Aaron/cromwell_tests/PDC000341/%40files/dbase/uniprot_human_jan2021_yeastENO1.fasta'
// fasta = '/home/ajm/code/nf-pdc/resources/uniprot_human_jan2021_yeastENO1_subset.fasta'
// spectral_library = 'https://panoramaweb.org/_webdav/MacCoss/Aaron/cromwell_tests/PDC000341/%40files/dbase/pan_human_library.dlib'
// spectral_library = '/home/ajm/code/nf-pdc/resources/uniprot_human_jan2021_yeastENO1_subset-z3_nce33-prosit.dlib'
// skyline_template_file = '/home/ajm/code/nf-pdc/nf-pdc/resources/template.sky.zip'
skyline_template_file = 'https://raw.githubusercontent.com/uw-maccosslab/nf-pdc-dia/master/resources/template_iRT.sky.zip'
gene_level_data = 'https://raw.githubusercontent.com/uw-maccosslab/nf-pdc-dia/master/resources/prhuman2gene-2023-05-24.csv'
panorama.skyline_folder = null
panorama.mzml_folder = null
panorama.reports_folder = null
skyline.protein_group_args = ''
skyline.save_intermediate_output = true
skyline.minimize = true
qc_report.precursor_report_template = 'https://raw.githubusercontent.com/uw-maccosslab/nf-pdc-dia/master/resources/precursor_quality.skyr'
qc_report.replicate_report_template = 'https://raw.githubusercontent.com/uw-maccosslab/nf-pdc-dia/master/resources/replicate_quality.skyr'
qc_report.standard_proteins = []
s3_upload.bucket_name = null
s3_upload.prefix_dir = null
s3_upload.access_key = null
}
docker {
enabled = true
}
/*
* Set up secrets in the environment.
* Need to do it this way because Nextflow doesn't allow the use of secrets when running on AWS Batch
*/
// ensure PanoramaWeb API keys will work. see: https://github.com/nextflow-io/nextflow/issues/3690
env.PLACEHOLDER = "PLACEHOLDER_VALUE"
secret_value = nextflow.secret.SecretsLoader.instance.load().getSecret("PANORAMA_API_KEY")
secret_value = nextflow.secret.SecretsLoader.instance.load().getSecret("S3_SECRET_ACCESS_KEY")
if(secret_value) {
env.PANORAMA_API_KEY = secret_value.value
env.S3_SECRET_ACCESS_KEY = secret_value.value
}
aws {
batch {
// NOTE: this setting is only required if the AWS CLI tool is installed in a custom AMI
cliPath = '/usr/local/aws-cli/v2/current/bin/aws'
logsGroup = '/batch/tei-nextflow-batch'
maxConnections = 20
connectionTimeout = 10000
uploadStorageClass = 'INTELLIGENT_TIERING'
storageEncryption = 'AES256'
retryMode = 'standard'
}
region = 'us-west-2'
}
// Execution Profiles
profiles {
/*
* Params for running pipeline on the local computer (e.g.:
* your laptop). These can be overridden in the local config file.
*/
standard {
process.executor = 'local'
// limit nextflow to running 1 task at a time
executor.queueSize = 8
params.max_memory = '32.GB'
params.max_cpus = 12
params.max_time = '240.h'
// where to cache mzml files after running msconvert
params.mzml_cache_directory = '~/code/nf-pdc/mzml_cache'
params.panorama_cache_directory = '~/code/nf-pdc/raw_cache'
}
aws {
process.executor = 'awsbatch'
process.queue = 'nextflow_basic_ec2'
// params for running pipeline on aws batch
// These can be overridden in local config file
// max params allowed for your AWS Batch compute environment
params.max_memory = 124.GB
params.max_cpus = 32
params.max_time = '240.h'
// where to cache mzml files after running msconvert
params.mzml_cache_directory = 's3://mc-tei-rex-nextflow-dda/dia/mzml_cache'
params.panorama_cache_directory = 's3://mc-tei-rex-nextflow-dda/panorama_cache'
}
slurm {
process.executor = 'slurm'
params.max_memory = '12.GB'
params.max_cpus = 8
params.max_time = '240.h'
// where to cache mzml files after running msconvert
params.mzml_cache_directory = '/data/mass_spec/nextflow/nf-teirex-dia/mzml_cache'
params.panorama_cache_directory = '/data/mass_spec/nextflow/panorama/raw_cache'
}
}
// Capture exit codes from upstream processes when piping
process.shell = ['/bin/bash', '-euo', 'pipefail']
def trace_timestamp = new java.util.Date().format( 'yyyy-MM-dd_HH-mm-ss')
timeline {
enabled = true
file = "${params.report_dir}/execution_timeline_${trace_timestamp}.html"
}
report {
enabled = true
file = "${params.report_dir}/execution_report_${trace_timestamp}.html"
}
trace {
enabled = true
file = "${params.report_dir}/execution_trace_${trace_timestamp}.txt"
}
dag {
enabled = false
file = "${params.report_dir}/pipeline_dag_${trace_timestamp}.html"
}
// Load base.config by default for all pipelines
includeConfig 'conf/base.config'