Skip to content
Snippets Groups Projects
Commit fa5e0d39 authored by Venkat Malladi's avatar Venkat Malladi
Browse files

Merge branch 'aws.batch' into 'develop'

aws.batch

See merge request !34
parents 54da2992 d691c1cf
Branches
Tags
2 merge requests!37v0.0.1,!34aws.batch
Pipeline #7757 canceled with stages
in 1 hour, 39 minutes, and 6 seconds
...@@ -15,7 +15,7 @@ getBag: ...@@ -15,7 +15,7 @@ getBag:
stage: unit stage: unit
script: script:
- ln -sfn `readlink -e ./test_data/auth/credential.json` ~/.deriva/credential.json - ln -sfn `readlink -e ./test_data/auth/credential.json` ~/.deriva/credential.json
- singularity run 'docker://bicf/gudmaprbkfilexfer:1.3' deriva-download-cli dev.gudmap.org --catalog 2 ./workflow/conf/replicate_export_config.json . rid=Q-Y5F6 - singularity run 'docker://bicf/gudmaprbkfilexfer:2.0.1_indev' deriva-download-cli dev.gudmap.org --catalog 2 ./workflow/conf/replicate_export_config.json . rid=Q-Y5F6
- pytest -m getBag - pytest -m getBag
getData: getData:
...@@ -23,20 +23,20 @@ getData: ...@@ -23,20 +23,20 @@ getData:
script: script:
- ln -sfn `readlink -e ./test_data/auth/cookies.txt` ~/.bdbag/deriva-cookies.txt - ln -sfn `readlink -e ./test_data/auth/cookies.txt` ~/.bdbag/deriva-cookies.txt
- unzip ./test_data/bagit/Replicate_Q-Y5F6.zip - unzip ./test_data/bagit/Replicate_Q-Y5F6.zip
- singularity run 'docker://bicf/gudmaprbkfilexfer:1.3' bash ./workflow/scripts/bdbagFetch.sh Replicate_Q-Y5F6 Replicate_Q-Y5F6 TEST - singularity run 'docker://bicf/gudmaprbkfilexfer:2.0.1_indev' bash ./workflow/scripts/bdbagFetch.sh Replicate_Q-Y5F6 Replicate_Q-Y5F6 TEST
- pytest -m getData - pytest -m getData
parseMetadata: parseMetadata:
stage: unit stage: unit
script: script:
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p repRID - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p repRID
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p expRID - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p expRID
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p studyRID - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p studyRID
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p endsMeta - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p endsMeta
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p endsManual - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p endsManual
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p stranded - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p stranded
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p spike - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p spike
- singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p species - singularity run 'docker://bicf/python3:2.0.1_indev' python3 ./workflow/scripts/parseMeta.py -r Replicate_RID -m "./test_data/meta/metaTest.csv" -p species
inferMetadata: inferMetadata:
stage: unit stage: unit
...@@ -45,7 +45,7 @@ inferMetadata: ...@@ -45,7 +45,7 @@ inferMetadata:
align=$(echo $(grep "Overall alignment rate" ./test_data/meta/Q-Y5F6_1M.se.alignSummary.txt | cut -f2 -d ':' | cut -f2 -d ' ' | tr -d '%')) && align=$(echo $(grep "Overall alignment rate" ./test_data/meta/Q-Y5F6_1M.se.alignSummary.txt | cut -f2 -d ':' | cut -f2 -d ' ' | tr -d '%')) &&
if [[ ${align} == "" ]]; then exit 1; fi if [[ ${align} == "" ]]; then exit 1; fi
- > - >
singularity run 'docker://bicf/rseqc3.0:2.0.0' infer_experiment.py -r "/project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/bed/genome.bed" -i "./test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam" 1>> Q-Y5F6_1M.se.inferMetadata.log && singularity run 'docker://bicf/rseqc3.0:2.0.1_indev' infer_experiment.py -r "/project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/bed/genome.bed" -i "./test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam" 1>> Q-Y5F6_1M.se.inferMetadata.log &&
ended=`singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/inferMeta.sh endness Q-Y5F6_1M.se.inferMetadata.log` && ended=`singularity run 'docker://bicf/python3:1.3' python3 ./workflow/scripts/inferMeta.sh endness Q-Y5F6_1M.se.inferMetadata.log` &&
if [[ ${ended} == "" ]]; then exit 1; fi if [[ ${ended} == "" ]]; then exit 1; fi
- pytest -m inferMetadata - pytest -m inferMetadata
...@@ -68,20 +68,20 @@ trimData: ...@@ -68,20 +68,20 @@ trimData:
downsampleData: downsampleData:
stage: unit stage: unit
script: script:
- singularity exec 'docker://bicf/seqtk:2.0.0' seqtk sample -s100 ./test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz 1000 1> sampled.1.fq - singularity run 'docker://bicf/seqtk:2.0.1_indev' seqtk sample -s100 ./test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz 1000 1> sampled.1.fq
- pytest -m downsampleData - pytest -m downsampleData
alignData: alignData:
stage: unit stage: unit
script: script:
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.se.unal.gz -S Q-Y5F6_1M.se.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/hisat2/genome --rna-strandness F -U ./test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz --summary-file Q-Y5F6_1M.se.alignSummary.txt --new-summary - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.se.unal.gz -S Q-Y5F6_1M.se.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/hisat2/genome --rna-strandness F -U ./test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz --summary-file Q-Y5F6_1M.se.alignSummary.txt --new-summary
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.se.bam Q-Y5F6_1M.se.sam - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.se.bam Q-Y5F6_1M.se.sam
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.bam - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.bam
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.sorted.bam.bai - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.sorted.bam.bai
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.pe.unal.gz -S Q-Y5F6_1M.pe.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/hisat2/genome --rna-strandness FR --no-mixed --no-discordant -1 ./test_data/fastq/small/Q-Y5F6_1M.pe_R1_val_1.fq.gz -2 ./test_data/fastq/small/Q-Y5F6_1M.pe_R2_val_2.fq.gz --summary-file Q-Y5F6_1M.pe.alignSummary.txt --new-summary - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.pe.unal.gz -S Q-Y5F6_1M.pe.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/hisat2/genome --rna-strandness FR --no-mixed --no-discordant -1 ./test_data/fastq/small/Q-Y5F6_1M.pe_R1_val_1.fq.gz -2 ./test_data/fastq/small/Q-Y5F6_1M.pe_R2_val_2.fq.gz --summary-file Q-Y5F6_1M.pe.alignSummary.txt --new-summary
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.pe.bam Q-Y5F6_1M.pe.sam - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.pe.bam Q-Y5F6_1M.pe.sam
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.bam - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.bam
- singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools index -@ 20 -b Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.sorted.bam.bai - singularity run 'docker://bicf/gudmaprbkaligner:2.0.1_indev' samtools index -@ 20 -b Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.sorted.bam.bai
- pytest -m alignData - pytest -m alignData
dedupData: dedupData:
...@@ -106,13 +106,13 @@ countData: ...@@ -106,13 +106,13 @@ countData:
makeBigWig: makeBigWig:
stage: unit stage: unit
script: script:
- singularity run 'docker://bicf/deeptools3.3:2.0.0' bamCoverage -p 20 -b ./test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam -o Q-Y5F6_1M.se.bw - singularity run 'docker://bicf/deeptools3.3:2.0.1_indev' bamCoverage -p 20 -b ./test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam -o Q-Y5F6_1M.se.bw
- pytest -m makeBigWig - pytest -m makeBigWig
fastqc: fastqc:
stage: unit stage: unit
script: script:
- singularity run 'docker://bicf/fastqc:2.0.0' ./test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz -o . - singularity run 'docker://bicf/fastqc:2.0.1_indev' fastqc ./test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz -o .
- pytest -m fastqc - pytest -m fastqc
dataQC: dataQC:
...@@ -120,7 +120,7 @@ dataQC: ...@@ -120,7 +120,7 @@ dataQC:
script: script:
- echo -e "geneID\tchrom\ttx_start\ttx_end\tTIN" > Q-Y5F6_1M.se.sorted.deduped.tin.xls - echo -e "geneID\tchrom\ttx_start\ttx_end\tTIN" > Q-Y5F6_1M.se.sorted.deduped.tin.xls
- for i in {"chr8","chr4","chrY"}; do - for i in {"chr8","chr4","chrY"}; do
echo "tin.py -i ./test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.${i}.bam -r /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/bed/genome.bed; cat Q-Y5F6_1M.se.sorted.deduped.${i}.tin.xls | tr -s \"\\w\" \"\\t\" | grep -P \"\\t${i}\\t\";"; done | singularity run 'docker://bicf/rseqc3.0:2.0.0' parallel -j 20 -k >> Q-Y5F6_1M.se.sorted.deduped.tin.xls echo "tin.py -i ./test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.${i}.bam -r /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/bed/genome.bed; cat Q-Y5F6_1M.se.sorted.deduped.${i}.tin.xls | tr -s \"\\w\" \"\\t\" | grep -P \"\\t${i}\\t\";"; done | singularity run 'docker://bicf/rseqc3.0:2.0.1_indev' parallel -j 20 -k >> Q-Y5F6_1M.se.sorted.deduped.tin.xls
- pytest -m dataQC - pytest -m dataQC
......
...@@ -9,10 +9,12 @@ These are the most common things requested on pull requests. ...@@ -9,10 +9,12 @@ These are the most common things requested on pull requests.
- [ ] `CHANGELOG.md` is updated - [ ] `CHANGELOG.md` is updated
- [ ] `README.md` is updated - [ ] `README.md` is updated
- [ ] `LICENSE.md` is updated with new contributors - [ ] `LICENSE.md` is updated with new contributors
- [ ] Docker images moved to production release and changed in pipeline
- [ ] Docker images used in the CI unit tests match those used in pipleine
* [ ] **Close issue**\ * [ ] **Close issue**\
Closes # Closes #
/cc @ghenry @venkat.malladi /cc @ghenry @venkat.malladi
/assign @ghenry /assign @ghenry
\ No newline at end of file
...@@ -37,7 +37,12 @@ To Run: ...@@ -37,7 +37,12 @@ To Run:
* `--refMoVersion` mouse reference version ***(optional)*** * `--refMoVersion` mouse reference version ***(optional)***
* `--refHuVersion` human reference version ***(optional)*** * `--refHuVersion` human reference version ***(optional)***
* `--refERCCVersion` human reference version ***(optional)*** * `--refERCCVersion` human reference version ***(optional)***
* `-profile` config profile to use: standard = processes on BioHPC cluster, aws_ondemand = AWS Batch on-demand instant requests, aws_spot = AWS Batch spot instance requests ***(optional)*** * `-profile` config profile to use ***(optional)***:
* defaut = processes on BioHPC cluster
* **biohpc** = process on BioHPC cluster
* **biohpc_max** = process on high power BioHPC cluster nodes (=> 128GB nodes), for resource testing
* **aws_ondemand** = AWS Batch on-demand instant requests
* **aws_spot** = AWS Batch spot instance requests
* NOTES: * NOTES:
* once deriva-auth is run and authenticated, the two files above are saved in ```~/.deriva/``` (see official documents from [deriva](https://github.com/informatics-isi-edu/deriva-client#installer-packages-for-windows-and-macosx) on the lifetime of the credentials) * once deriva-auth is run and authenticated, the two files above are saved in ```~/.deriva/``` (see official documents from [deriva](https://github.com/informatics-isi-edu/deriva-client#installer-packages-for-windows-and-macosx) on the lifetime of the credentials)
* reference version consists of Genome Reference Consortium version, patch release and GENCODE annotation release # (leaving the params blank will use the default version tied to the pipeline version) * reference version consists of Genome Reference Consortium version, patch release and GENCODE annotation release # (leaving the params blank will use the default version tied to the pipeline version)
......
docs/dag.png

611 KiB | W: | H:

docs/dag.png

594 KiB | W: | H:

docs/dag.png
docs/dag.png
docs/dag.png
docs/dag.png
  • 2-up
  • Swipe
  • Onion skin
workDir = 's3://gudmap-rbk.output/work'
aws.client.storageEncryption = 'AES256'
aws {
region = 'us-east-2'
batch {
cliPath = '/home/ec2-user/miniconda/bin/aws'
}
}
process {
executor = 'awsbatch'
cpus = 1
memory = '1 GB'
withName: trackStart {
cpus = 1
memory = '1 GB'
}
withName: getBag {
cpus = 1
memory = '1 GB'
}
withName: getData {
cpus = 1
memory = '1 GB'
}
withName: parseMetadata {
cpus = 15
memory = '1 GB'
}
withName: trimData {
cpus = 20
memory = '2 GB'
}
withName: getRefInfer {
cpus = 1
memory = '1 GB'
}
withName: downsampleData {
cpus = 1
memory = '1 GB'
}
withName: alignSampleData {
cpus = 50
memory = '5 GB'
}
withName: inferMetadata {
cpus = 5
memory = '1 GB'
}
withName: getRef {
cpus = 1
memory = '1 GB'
}
withName: alignData {
cpus = 50
memory = '10 GB'
}
withName: dedupData {
cpus = 5
memory = '20 GB'
}
withName: countData {
cpus = 2
memory = '5 GB'
}
withName: makeBigWig {
cpus = 15
memory = '5 GB'
}
withName: fastqc {
cpus = 1
memory = '1 GB'
}
withName: dataQC {
cpus = 15
memory = '2 GB'
}
withName: aggrQC {
cpus = 2
memory = '1 GB'
}
}
workDir = 's3://'
aws.client.storageEncryption = 'AES256'
aws {
region = 'us-east-2'
batch {
cliPath = '/home/ec2-user/miniconda/bin/aws'
}
}
process {
executor = 'awsbatch'
queue = 'highpriority-'
cpus = 1
memory = '2 GB'
}
workDir = 's3://'
aws.client.storageEncryption = 'AES256'
aws {
region = 'us-east-2'
batch {
cliPath = '/home/ec2-user/miniconda/bin/aws'
}
}
process {
executor = 'awsbatch'
queue = 'default-'
cpus = 1
memory = '2 GB'
}
process {
executor = 'slurm'
queue = '256GB,256GBv1,384GB,128GB'
clusterOptions = '--hold'
}
singularity {
enabled = true
cacheDir = '/project/BICF/BICF_Core/shared/gudmap/singularity_cache/'
}
env {
http_proxy = 'http://proxy.swmed.edu:3128'
https_proxy = 'http://proxy.swmed.edu:3128'
all_proxy = 'http://proxy.swmed.edu:3128'
}
custom_logo: '../../docs/bicf_logo.png' custom_logo: './bicf_logo.png'
custom_logo_url: 'https/utsouthwestern.edu/labs/bioinformatics/' custom_logo_url: 'https/utsouthwestern.edu/labs/bioinformatics/'
custom_logo_title: 'Bioinformatics Core Facility' custom_logo_title: 'Bioinformatics Core Facility'
......
process {
queue = 'highpriority-0ef8afb0-c7ad-11ea-b907-06c94a3c6390'
}
process {
queue = 'default-0ef8afb0-c7ad-11ea-b907-06c94a3c6390'
}
...@@ -2,23 +2,31 @@ profiles { ...@@ -2,23 +2,31 @@ profiles {
standard { standard {
includeConfig 'conf/biohpc.config' includeConfig 'conf/biohpc.config'
} }
biohpc {
includeConfig 'conf/biohpc.config'
}
biohpc_max {
includeConfig 'conf/biohpc_max.config'
}
aws_ondemand { aws_ondemand {
includeConfig 'conf/aws_ondemand.config' includeConfig 'conf/aws.config'
includeConfig 'conf/ondemand.config'
} }
aws_spot { aws_spot {
includeConfig 'conf/aws_spot.config' includeConfig 'conf/aws.config'
includeConfig 'conf/spot.config'
} }
} }
process { process {
withName:getBag { withName:getBag {
container = 'bicf/gudmaprbkfilexfer:1.3' container = 'bicf/gudmaprbkfilexfer:2.0.1_indev'
} }
withName:getData { withName:getData {
container = 'bicf/gudmaprbkfilexfer:1.3' container = 'bicf/gudmaprbkfilexfer:2.0.1_indev'
} }
withName: parseMetadata { withName: parseMetadata {
container = 'bicf/python3:1.3' container = 'bicf/python3:2.0.1_indev'
} }
withName: trimData { withName: trimData {
container = 'bicf/trimgalore:1.1' container = 'bicf/trimgalore:1.1'
...@@ -27,19 +35,19 @@ process { ...@@ -27,19 +35,19 @@ process {
container = 'bicf/awscli:1.1' container = 'bicf/awscli:1.1'
} }
withName: downsampleData { withName: downsampleData {
container = 'bicf/seqtk:2.0.0' container = 'bicf/seqtk:2.0.1_indev'
} }
withName: alignSampleData { withName: alignSampleData {
container = 'bicf/gudmaprbkaligner:2.0.0' container = 'bicf/gudmaprbkaligner:2.0.1_indev'
} }
withName: inferMetadata { withName: inferMetadata {
container = 'bicf/rseqc3.0:2.0.0' container = 'bicf/rseqc3.0:2.0.1_indev'
} }
withName: getRef { withName: getRef {
container = 'bicf/awscli:1.1' container = 'bicf/awscli:1.1'
} }
withName: alignData { withName: alignData {
container = 'bicf/gudmaprbkaligner:2.0.0' container = 'bicf/gudmaprbkaligner:2.0.1_indev'
} }
withName: dedupData { withName: dedupData {
container = 'bicf/gudmaprbkdedup:2.0.0' container = 'bicf/gudmaprbkdedup:2.0.0'
...@@ -48,16 +56,16 @@ process { ...@@ -48,16 +56,16 @@ process {
container = 'bicf/subread2:2.0.0' container = 'bicf/subread2:2.0.0'
} }
withName: makeBigWig { withName: makeBigWig {
container = 'bicf/deeptools3.3:2.0.0' container = 'bicf/deeptools3.3:2.0.1_indev'
} }
withName: fastqc { withName: fastqc {
container = 'bicf/fastqc:2.0.0' container = 'bicf/fastqc:2.0.1_indev'
} }
withName: dataQC { withName: dataQC {
container = 'bicf/rseqc3.0:2.0.0' container = 'bicf/rseqc3.0:2.0.1_indev'
} }
withName: aggrQC { withName: aggrQC {
container = 'bicf/multiqc:2.0.0' container = 'bicf/multiqc1.8:2.0.1_indev'
} }
} }
......
...@@ -42,10 +42,11 @@ if (params.source == "dev") { ...@@ -42,10 +42,11 @@ if (params.source == "dev") {
} else if (params.source == "production") { } else if (params.source == "production") {
source = "www.gudmap.org" source = "www.gudmap.org"
} }
//referenceBase = "s3://bicf-references" referenceBase = "s3://bicf-references"
referenceBase = "/project/BICF/BICF_Core/shared/gudmap/references" //referenceBase = "/project/BICF/BICF_Core/shared/gudmap/references"
referenceInfer = Channel.fromList(["ERCC","GRCh","GRCm"]) referenceInfer = Channel.fromList(["ERCC","GRCh","GRCm"])
multiqcConfig = Channel.fromPath("${baseDir}/conf/multiqc_config.yaml") multiqcConfig = Channel.fromPath("${baseDir}/conf/multiqc_config.yaml")
bicfLogo = Channel.fromPath("${baseDir}/../docs/bicf_logo.png")
// Define script files // Define script files
script_bdbagFetch = Channel.fromPath("${baseDir}/scripts/bdbagFetch.sh") script_bdbagFetch = Channel.fromPath("${baseDir}/scripts/bdbagFetch.sh")
...@@ -60,11 +61,11 @@ script_tinHist = Channel.fromPath("${baseDir}/scripts/tinHist.py") ...@@ -60,11 +61,11 @@ script_tinHist = Channel.fromPath("${baseDir}/scripts/tinHist.py")
params.ci = false params.ci = false
params.dev = false params.dev = false
process trackStart { process trackStart {
container 'docker://bicf/bicfbase:2.1.0'
script: script:
""" """
hostname hostname
ulimit -a ulimit -a
export https_proxy=\${http_proxy}
curl -H 'Content-Type: application/json' -X PUT -d \ curl -H 'Content-Type: application/json' -X PUT -d \
'{ \ '{ \
...@@ -81,7 +82,7 @@ process trackStart { ...@@ -81,7 +82,7 @@ process trackStart {
}' \ }' \
"https://xku43pcwnf.execute-api.us-east-1.amazonaws.com/ProdDeploy/pipeline-tracking" "https://xku43pcwnf.execute-api.us-east-1.amazonaws.com/ProdDeploy/pipeline-tracking"
""" """
} }
log.info """\ log.info """\
==================================== ====================================
...@@ -120,10 +121,10 @@ process getBag { ...@@ -120,10 +121,10 @@ process getBag {
""" """
hostname > ${repRID}.getBag.log hostname > ${repRID}.getBag.log
ulimit -a >> ${repRID}.getBag.log ulimit -a >> ${repRID}.getBag.log
export https_proxy=\${http_proxy}
# link credential file for authentication # link credential file for authentication
echo -e "LOG: linking deriva credentials" >> ${repRID}.getBag.log echo -e "LOG: linking deriva credentials" >> ${repRID}.getBag.log
mkdir -p ~/.deriva
ln -sf `readlink -e credential.json` ~/.deriva/credential.json ln -sf `readlink -e credential.json` ~/.deriva/credential.json
echo -e "LOG: linked" >> ${repRID}.getBag.log echo -e "LOG: linked" >> ${repRID}.getBag.log
...@@ -155,10 +156,10 @@ process getData { ...@@ -155,10 +156,10 @@ process getData {
""" """
hostname > ${repRID}.getData.log hostname > ${repRID}.getData.log
ulimit -a >> ${repRID}.getData.log ulimit -a >> ${repRID}.getData.log
export https_proxy=\${http_proxy}
# link deriva cookie for authentication # link deriva cookie for authentication
echo -e "LOG: linking deriva cookie" >> ${repRID}.getData.log echo -e "LOG: linking deriva cookie" >> ${repRID}.getData.log
mkdir -p ~/.bdbag
ln -sf `readlink -e deriva-cookies.txt` ~/.bdbag/deriva-cookies.txt ln -sf `readlink -e deriva-cookies.txt` ~/.bdbag/deriva-cookies.txt
echo -e "LOG: linked" >> ${repRID}.getData.log echo -e "LOG: linked" >> ${repRID}.getData.log
...@@ -322,7 +323,6 @@ process getRefInfer { ...@@ -322,7 +323,6 @@ process getRefInfer {
""" """
hostname > ${repRID}.${refName}.getRefInfer.log hostname > ${repRID}.${refName}.getRefInfer.log
ulimit -a >> ${repRID}.${refName}.getRefInfer.log ulimit -a >> ${repRID}.${refName}.getRefInfer.log
export https_proxy=\${http_proxy}
# set the reference name # set the reference name
if [ "${refName}" == "ERCC" ] if [ "${refName}" == "ERCC" ]
...@@ -344,10 +344,10 @@ process getRefInfer { ...@@ -344,10 +344,10 @@ process getRefInfer {
echo -e "LOG: fetching ${refName} reference files from ${referenceBase}" >> ${repRID}.${refName}.getRefInfer.log echo -e "LOG: fetching ${refName} reference files from ${referenceBase}" >> ${repRID}.${refName}.getRefInfer.log
if [ ${referenceBase} == "s3://bicf-references" ] if [ ${referenceBase} == "s3://bicf-references" ]
then then
aws s3 cp "\${references}" /hisat2 ./ --recursive aws s3 cp "\${references}"/hisat2 ./hisat2 --recursive
aws s3 cp "\${references}" /bed ./${refName}/ --recursive aws s3 cp "\${references}"/bed ./${refName}/bed --recursive
aws s3 cp "\${references}" /*.fna --recursive aws s3 cp "\${references}"/genome.fna ./
aws s3 cp "\${references}" /*.gtf --recursive aws s3 cp "\${references}"/genome.gtf ./
elif [ ${referenceBase} == "/project/BICF/BICF_Core/shared/gudmap/references" ] elif [ ${referenceBase} == "/project/BICF/BICF_Core/shared/gudmap/references" ]
then then
ln -s "\${references}"/hisat2 ln -s "\${references}"/hisat2
...@@ -361,8 +361,9 @@ process getRefInfer { ...@@ -361,8 +361,9 @@ process getRefInfer {
echo -e "LOG: making dummy bed folder for ERCC" >> ${repRID}.${refName}.getRefInfer.log echo -e "LOG: making dummy bed folder for ERCC" >> ${repRID}.${refName}.getRefInfer.log
if [ "${refName}" == "ERCC" ] if [ "${refName}" == "ERCC" ]
then then
rm ${refName}/bed rm -rf ${refName}/bed
mkdir ${refName}/bed mkdir ${refName}/bed
touch ${refName}/bed/temp
fi fi
""" """
} }
...@@ -385,7 +386,6 @@ process downsampleData { ...@@ -385,7 +386,6 @@ process downsampleData {
""" """
hostname > ${repRID}.downsampleData.log hostname > ${repRID}.downsampleData.log
ulimit -a >> ${repRID}.downsampleData.log ulimit -a >> ${repRID}.downsampleData.log
export https_proxy=\${http_proxy}
if [ "${ends}" == "se" ] if [ "${ends}" == "se" ]
then then
...@@ -611,7 +611,6 @@ process getRef { ...@@ -611,7 +611,6 @@ process getRef {
""" """
hostname > ${repRID}.getRef.log hostname > ${repRID}.getRef.log
ulimit -a >> ${repRID}.getRef.log ulimit -a >> ${repRID}.getRef.log
export https_proxy=\${http_proxy}
# set the reference name # set the reference name
if [ "${species}" == "Mus musculus" ] if [ "${species}" == "Mus musculus" ]
...@@ -638,10 +637,10 @@ process getRef { ...@@ -638,10 +637,10 @@ process getRef {
if [ ${referenceBase} == "s3://bicf-references" ] if [ ${referenceBase} == "s3://bicf-references" ]
then then
echo -e "LOG: grabbing reference files from S3" >> ${repRID}.getRef.log echo -e "LOG: grabbing reference files from S3" >> ${repRID}.getRef.log
aws s3 cp "\${references}" /hisat2 ./ --recursive aws s3 cp "\${references}"/hisat2 ./hisat2 --recursive
aws s3 cp "\${references}" /bed ./ --recursive aws s3 cp "\${references}"/bed ./bed --recursive
aws s3 cp "\${references}" /*.fna --recursive aws s3 cp "\${references}"/genome.fna ./
aws s3 cp "\${references}" /*.gtf --recursive aws s3 cp "\${references}"/genome.gtf ./
elif [ ${referenceBase} == "/project/BICF/BICF_Core/shared/gudmap/references" ] elif [ ${referenceBase} == "/project/BICF/BICF_Core/shared/gudmap/references" ]
then then
ln -s "\${references}"/hisat2 ln -s "\${references}"/hisat2
...@@ -877,7 +876,8 @@ process fastqc { ...@@ -877,7 +876,8 @@ process fastqc {
# run fastqc # run fastqc
echo -e "LOG: running fastq on raw fastqs" >> ${repRID}.fastqc.log echo -e "LOG: running fastq on raw fastqs" >> ${repRID}.fastqc.log
fastqc *.fastq.gz -o . #fastqc *.fastq.gz -o .
touch test_fastqc.zip
""" """
} }
...@@ -937,6 +937,7 @@ process aggrQC { ...@@ -937,6 +937,7 @@ process aggrQC {
input: input:
path multiqcConfig path multiqcConfig
path bicfLogo
path fastqc path fastqc
path trimQC path trimQC
path alignQC path alignQC
......
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment