Skip to content
Snippets Groups Projects
Commit ad5bf09e authored by Gervaise Henry's avatar Gervaise Henry :cowboy:
Browse files

Merge branch '11-deriva.upload' into 'develop'

Resolve "process_derivaUpload"

Closes #24, #75, and #11

See merge request !53
parents 30143e2f 5e6b9051
Branches
Tags
2 merge requests!58Develop,!53Resolve "process_derivaUpload"
Pipeline #8733 failed with stages
in 8 minutes and 17 seconds
Showing
with 1035 additions and 265 deletions
This diff is collapsed.
...@@ -5,9 +5,9 @@ These are the most common things requested on pull requests. ...@@ -5,9 +5,9 @@ These are the most common things requested on pull requests.
- [ ] This comment contains a description of changes (with reason) - [ ] This comment contains a description of changes (with reason)
- [ ] If you've fixed a bug or added code that should be tested, add tests! - [ ] If you've fixed a bug or added code that should be tested, add tests!
- [ ] Documentation in `docs` is updated - [ ] Documentation in `docs` is updated
- [ ] Replace dag.png with the most recent CI pipleine integrated_pe artifact - [ ] Replace dag.png with the most recent CI pipeline integrated_pe artifact
- [ ] Replace software_versions_mqc.yaml with the most recent CI pipleine generateVersions artifact - [ ] Replace software_versions_mqc.yaml with the most recent CI pipeline generateVersions artifact
- [ ] Replace software_references_mqc.yaml with the most recent CI pipleine generateVersions artifact - [ ] Replace software_references_mqc.yaml with the most recent CI pipeline generateVersions artifact
- [ ] `CHANGELOG.md` is updated - [ ] `CHANGELOG.md` is updated
- [ ] `README.md` is updated - [ ] `README.md` is updated
- [ ] `LICENSE.md` is updated with new contributors - [ ] `LICENSE.md` is updated with new contributors
......
# v0.0.4 (in development) # v0.1.0 (in development)
**User Facing** **User Facing**
* Add option to pull references from datahub * Add option to pull references from datahub
* Add option to send email on workflow error, with pipeline error message * Add option to send email on workflow error, with pipeline error message
* Add versions and paper references of software used to report * Add versions and paper references of software used to report
* Upload input bag
* Upload execution run
* Upload mRNA QC
* Create and upload output bag
* Add optional to not upload
**Background** **Background**
* Remove (comment out) option to pull references from S3 * Remove (comment out) option to pull references from S3
...@@ -10,11 +15,11 @@ ...@@ -10,11 +15,11 @@
* Start using new gudmaprbk dockerhub (images autobuilt) * Start using new gudmaprbk dockerhub (images autobuilt)
* Moved consistency checks to be fully python * Moved consistency checks to be fully python
* Changed order of steps so that fastqc is done after the trim step * Changed order of steps so that fastqc is done after the trim step
* Change docker images to production
* Add automated version badges
*Known Bugs* *Known Bugs*
* Datahub reference pull uses dev.gudmap.org as source until referencencs are placed on production * Datahub reference pull uses dev.gudmap.org as source until referencencs are placed on production
* outputBag does not contain fetch for processed data
* Does not include automatic data upload
* Override params (inputBag, fastq, species) aren't checked for integrity * Override params (inputBag, fastq, species) aren't checked for integrity
<hr> <hr>
......
...@@ -37,9 +37,12 @@ To Run: ...@@ -37,9 +37,12 @@ To Run:
* **dev** = [dev.gudmap.org](dev.gudmap.org) (default, does not contain all data) * **dev** = [dev.gudmap.org](dev.gudmap.org) (default, does not contain all data)
* **staging** = [staging.gudmap.org](staging.gudmap.org) (does not contain all data) * **staging** = [staging.gudmap.org](staging.gudmap.org) (does not contain all data)
* **production** = [www.gudmap.org](www.gudmap.org) (***does contain all data***) * **production** = [www.gudmap.org](www.gudmap.org) (***does contain all data***)
* `--refMoVersion` mouse reference version ***(optional)*** * `--refMoVersion` mouse reference version ***(optional, default = 38.p6.vM22)***
* `--refHuVersion` human reference version ***(optional)*** * `--refHuVersion` human reference version ***(optional, default = 38.p12.v31)***
* `--refERCCVersion` human reference version ***(optional)*** * `--refERCCVersion` human reference version ***(optional, default = 92)***
* `--upload` option to not upload output back to the data-hub ***(optional, default = true)***
* **true** = upload outputs to the data-hub
* **false** = do *NOT* upload outputs to the data-hub
* `-profile` config profile to use ***(optional)***: * `-profile` config profile to use ***(optional)***:
* defaut = processes on BioHPC cluster * defaut = processes on BioHPC cluster
* **biohpc** = process on BioHPC cluster * **biohpc** = process on BioHPC cluster
...@@ -47,7 +50,7 @@ To Run: ...@@ -47,7 +50,7 @@ To Run:
* **aws_ondemand** = AWS Batch on-demand instant requests * **aws_ondemand** = AWS Batch on-demand instant requests
* **aws_spot** = AWS Batch spot instance requests * **aws_spot** = AWS Batch spot instance requests
* `--email` email address(es) to send failure notification (comma separated) ***(optional)***: * `--email` email address(es) to send failure notification (comma separated) ***(optional)***:
* e.g: `--email 'venkat.malladi@utsouthwestern.edu,Gervaise.Henry@UTSouthwestern.edu'` * e.g: `--email 'Venkat.Malladi@utsouthwestern.edu,Gervaise.Henry@UTSouthwestern.edu'`
* NOTES: * NOTES:
* once deriva-auth is run and authenticated, the two files above are saved in ```~/.deriva/``` (see official documents from [deriva](https://github.com/informatics-isi-edu/deriva-client#installer-packages-for-windows-and-macosx) on the lifetime of the credentials) * once deriva-auth is run and authenticated, the two files above are saved in ```~/.deriva/``` (see official documents from [deriva](https://github.com/informatics-isi-edu/deriva-client#installer-packages-for-windows-and-macosx) on the lifetime of the credentials)
......
docs/dag.png

762 KiB | W: | H:

docs/dag.png

1.36 MiB | W: | H:

docs/dag.png
docs/dag.png
docs/dag.png
docs/dag.png
  • 2-up
  • Swipe
  • Onion skin
...@@ -4,7 +4,7 @@ ...@@ -4,7 +4,7 @@
description: 'This section describes references for the tools used.' description: 'This section describes references for the tools used.'
plot_type: 'html' plot_type: 'html'
data: | data: |
<h3 id="references">References</h3> <h3 id="references">References</h3>
<ol style="list-style-type: decimal"> <ol style="list-style-type: decimal">
<li><strong>python</strong>:</li> <li><strong>python</strong>:</li>
...@@ -41,7 +41,7 @@ ...@@ -41,7 +41,7 @@
<li><strong>hisat2</strong>:</li> <li><strong>hisat2</strong>:</li>
</ol> </ol>
<ul> <ul>
<li>Kim ,D.,Paggi, J.M., Park, C., Bennett, C., Salzberg, S.L. Graph-based genome alignment and genotyping with HISAT2 and HISAT-genotype. 2019 Nat Biotechnol. 2019 Aug;37(8):907-915. doi:<a href="https://doi.org/10.1038/s41587-019-0201-4">10.1038/s41587-019-0201-4</a></li> <li>Kim ,D.,Paggi, J.M., Park, C., Bennett, C., Salzberg, S.L. 2019 Graph-based genome alignment and genotyping with HISAT2 and HISAT-genotype. Nat Biotechnol. Aug;37(8):907-915. doi:<a href="https://doi.org/10.1038/s41587-019-0201-4">10.1038/s41587-019-0201-4</a>.</li>
</ul> </ul>
<ol start="7" style="list-style-type: decimal"> <ol start="7" style="list-style-type: decimal">
<li><strong>samtools</strong>:</li> <li><strong>samtools</strong>:</li>
......
...@@ -6,19 +6,19 @@ ...@@ -6,19 +6,19 @@
description: 'are collected for pipeline version.' description: 'are collected for pipeline version.'
data: | data: |
<dl class="dl-horizontal"> <dl class="dl-horizontal">
<dt>Python</dt><dd>v3.7.7</dd> <dt>Python</dt><dd>v3.8.3</dd>
<dt>DERIVA</dt><dd>v1.0.0</dd> <dt>DERIVA</dt><dd>v1.3.0</dd>
<dt>BDBag</dt><dd>v1.5.6</dd> <dt>BDBag</dt><dd>v1.5.6</dd>
<dt>RSeQC</dt><dd>v3.0.1</dd> <dt>RSeQC</dt><dd>v4.0.0</dd>
<dt>Trim Galore!</dt><dd>v0.6.4</dd> <dt>Trim Galore!</dt><dd>v0.6.4_dev</dd>
<dt>HISAT2</dt><dd>v2.1.0</dd> <dt>HISAT2</dt><dd>v2.2.1</dd>
<dt>Samtools</dt><dd>v1.9</dd> <dt>Samtools</dt><dd>v1.11</dd>
<dt>picard (MarkDuplicates)</dt><dd>v2.23.0-SNAPSHOT</dd> <dt>picard (MarkDuplicates)</dt><dd>v2.23.9</dd>
<dt>featureCounts</dt><dd>v2.0.0</dd> <dt>featureCounts</dt><dd>v2.0.1</dd>
<dt>R</dt><dd>v3.6.3</dd> <dt>R</dt><dd>v4.0.3</dd>
<dt>deepTools</dt><dd>v3.3.2</dd> <dt>deepTools</dt><dd>v3.5.0</dd>
<dt>FastQC</dt><dd>v0.11.9</dd> <dt>FastQC</dt><dd>v0.11.9</dd>
<dt>MultiQC</dt><dd>v1.8</dd> <dt>MultiQC</dt><dd>v1.9</dd>
<dt>Pipeline Version</dt><dd>v0.0.4_indev</dd> <dt>Pipeline Version</dt><dd>v0.0.4_indev</dd>
</dl> </dl>
...@@ -5,52 +5,54 @@ ...@@ -5,52 +5,54 @@
module load singularity/3.5.3 module load singularity/3.5.3
module load pigz/2.4 module load pigz/2.4
ln -sfn /project/BICF/BICF_Core/shared/gudmap/test_data/* ../test_data/
mkdir -p NEW_test_data mkdir -p NEW_test_data
ln -sfn `readlink -e ./test_data/auth/credential.json` ~/.deriva/credential.json ln -sfn ./test_data/auth/credential.json ~/.deriva/credential.json
mkdir -p ./NEW_test_data/bag mkdir -p ./NEW_test_data/bag
singularity run 'docker://bicf/gudmaprbkfilexfer:1.3' deriva-download-cli dev.gudmap.org --catalog 2 ./workflow/conf/replicate_export_config.json . rid=Q-Y5F6 singularity run 'docker://gudmaprbk/deriva1.3:1.0.0' deriva-download-cli staging.gudmap.org --catalog 2 ../workflow/conf/Replicate_For_Input_Bag.json . rid=Q-Y5F6
cp Replicate_Q-Y5F6.zip ./NEW_test_data/bag/Replicate_Q-Y5F6.zip cp Q-Y5F6_inputBag.zip ./NEW_test_data/bag/Q-Y5F6_inputBag_xxxxxxxx.zip
mkdir -p ./NEW_test_data/fastq mkdir -p ./NEW_test_data/fastq
unzip ./test_data/bag/Replicate_Q-Y5F6.zip unzip ./NEW_test_data/bag/Q-Y5F6_inputBag_xxxxxxxx.zip
singularity run 'docker://bicf/gudmaprbkfilexfer:1.3' bash ./workflow/scripts/bdbagFetch.sh Replicate_Q-Y5F6 Replicate_Q-Y5F6 singularity run 'docker://gudmaprbk/deriva1.3:1.0.0' bash ../workflow/scripts/bdbagFetch.sh Q-Y5F6_inputBag Q-Y5F6
cp Replicate_Q-Y5F6.R1.fastq.gz ./NEW_test_data/fastq/Replicate_Q-Y5F6.R1.fastq.gz cp Q-Y5F6.R1.fastq.gz ./NEW_test_data/fastq/Q-Y5F6.R1.fastq.gz
cp Replicate_Q-Y5F6.R2.fastq.gz ./NEW_test_data/fastq/Replicate_Q-Y5F6.R2.fastq.gz cp Q-Y5F6.R2.fastq.gz ./NEW_test_data/fastq/Q-Y5F6.R2.fastq.gz
mkdir -p ./NEW_test_data/fastq/small mkdir -p ./NEW_test_data/fastq/small
singularity exec 'docker://bicf/seqtk:2.0.0' seqtk sample -s100 ./NEW_test_data/fastq/Replicate_Q-Y5F6.R1.fastq.gz 1000000 1> Q-Y5F6_1M.R1.fastq singularity exec 'docker://gudmaprbk/seqtk1.3:1.0.0' seqtk sample -s100 ./NEW_test_data/fastq/Q-Y5F6.R1.fastq.gz 1000000 1> Q-Y5F6_1M.R1.fastq
singularity exec 'docker://bicf/seqtk:2.0.0' seqtk sample -s100 ./NEW_test_data/fastq/Replicate_Q-Y5F6.R2.fastq.gz 1000000 1> Q-Y5F6_1M.R2.fastq singularity exec 'docker://gudmaprbk/seqtk1.3:1.0.0' seqtk sample -s100 ./NEW_test_data/fastq/Q-Y5F6.R2.fastq.gz 1000000 1> Q-Y5F6_1M.R2.fastq
pigz Q-Y5F6_1M.R1.fastq pigz Q-Y5F6_1M.R1.fastq
pigz Q-Y5F6_1M.R2.fastq pigz Q-Y5F6_1M.R2.fastq
cp Q-Y5F6_1M.R1.fastq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz cp Q-Y5F6_1M.R1.fastq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz
cp Q-Y5F6_1M.R2.fastq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.R2.fastq.gz cp Q-Y5F6_1M.R2.fastq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.R2.fastq.gz
mkdir -p ./NEW_test_data/meta mkdir -p ./NEW_test_data/meta
singularity run 'docker://bicf/trimgalore:1.1' trim_galore --gzip -q 25 --illumina --length 35 --basename Q-Y5F6_1M.se -j 20 ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz singularity run 'docker://gudmaprbk/trimgalore0.6.5:1.0.0' trim_galore --gzip -q 25 --illumina --length 35 --basename Q-Y5F6_1M.se -j 20 ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz
singularity run 'docker://bicf/trimgalore:1.1' trim_galore --gzip -q 25 --illumina --length 35 --paired --basename Q-Y5F6_1M.pe -j 20 ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.R2.fastq.gz singularity run 'docker://gudmaprbk/trimgalore0.6.5:1.0.0' trim_galore --gzip -q 25 --illumina --length 35 --paired --basename Q-Y5F6_1M.pe -j 20 ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.R2.fastq.gz
cp Q-Y5F6_1M.se_trimmed.fq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz cp Q-Y5F6_1M.se_trimmed.fq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz
cp Q-Y5F6_1M.pe_R1_val_1.fq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_R1_val_1.fq.gz cp Q-Y5F6_1M.pe_val_1.fq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_val_1.fq.gz
cp Q-Y5F6_1M.pe_R2_val_2.fq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_R2_val_2.fq.gz cp Q-Y5F6_1M.pe_val_2.fq.gz ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_val_2.fq.gz
cp Q-Y5F6_1M.R1.fastq.gz_trimming_report.txt ./NEW_test_data/meta/Q-Y5F6_1M.R1.fastq.gz_trimming_report.txt cp Q-Y5F6_1M.R1.fastq.gz_trimming_report.txt ./NEW_test_data/meta/Q-Y5F6_1M.R1.fastq.gz_trimming_report.txt
cp Q-Y5F6_1M.R2.fastq.gz_trimming_report.txt ./NEW_test_data/meta/Q-Y5F6_1M.R2.fastq.gz_trimming_report.txt cp Q-Y5F6_1M.R2.fastq.gz_trimming_report.txt ./NEW_test_data/meta/Q-Y5F6_1M.R2.fastq.gz_trimming_report.txt
touch metaTest.csv touch metaTest.csv
echo 'Replicate_RID,Experiment_RID,Study_RID,Paired_End,File_Type,Has_Strand_Specific_Information,Used_Spike_Ins,Species' > metaTest.csv echo 'Replicate_RID,Experiment_RID,Study_RID,Paired_End,File_Type,Has_Strand_Specific_Information,Used_Spike_Ins,Species,Read_Length' > metaTest.csv
echo 'Replicate_RID,Experiment_RID,Study_RID,uk,FastQ,no,no,Homo sapiens' >> metaTest.csv echo 'Replicate_RID,Experiment_RID,Study_RID,uk,FastQ,no,no,Homo sapiens,75' >> metaTest.csv
cp metaTest.csv ./NEW_test_data/meta/metaTest.csv cp metaTest.csv ./NEW_test_data/meta/metaTest.csv
mkdir -p ./NEW_test_data/bam mkdir -p ./NEW_test_data/bam
mkdir -p ./NEW_test_data/bam/small mkdir -p ./NEW_test_data/bam/small
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.se.unal.gz -S Q-Y5F6_1M.se.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/hisat2/genome --rna-strandness F -U ./NEW_test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz --summary-file Q-Y5F6_1M.se.alignSummary.txt --new-summary singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.se.unal.gz -S Q-Y5F6_1M.se.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/hisat2/genome --rna-strandness F -U ./NEW_test_data/fastq/small/Q-Y5F6_1M.se_trimmed.fq.gz --summary-file Q-Y5F6_1M.se.alignSummary.txt --new-summary
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.se.bam Q-Y5F6_1M.se.sam singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.se.bam Q-Y5F6_1M.se.sam
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.bam singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.bam
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.sorted.bam.bai singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.bam Q-Y5F6_1M.se.sorted.bam.bai
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.pe.unal.gz -S Q-Y5F6_1M.pe.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/hisat2/genome --rna-strandness FR --no-mixed --no-discordant -1 ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_R1_val_1.fq.gz -2 ./test_data/fastq/small/Q-Y5F6_1M.pe_R2_val_2.fq.gz --summary-file Q-Y5F6_1M.pe.alignSummary.txt --new-summary singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' hisat2 -p 20 --add-chrname --un-gz Q-Y5F6_1M.pe.unal.gz -S Q-Y5F6_1M.pe.sam -x /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/hisat2/genome --rna-strandness FR --no-mixed --no-discordant -1 ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_val_1.fq.gz -2 ./NEW_test_data/fastq/small/Q-Y5F6_1M.pe_val_2.fq.gz --summary-file Q-Y5F6_1M.pe.alignSummary.txt --new-summary
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.pe.bam Q-Y5F6_1M.pe.sam singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' samtools view -1 -@ 20 -F 4 -F 8 -F 256 -o Q-Y5F6_1M.pe.bam Q-Y5F6_1M.pe.sam
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.bam singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.bam
singularity run 'docker://bicf/gudmaprbkaligner:2.0.0' samtools index -@ 20 -b Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.sorted.bam.bai singularity run 'docker://gudmaprbk/hisat2.2.1:1.0.0' samtools index -@ 20 -b Q-Y5F6_1M.pe.sorted.bam Q-Y5F6_1M.pe.sorted.bam.bai
cp Q-Y5F6_1M.se.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.bam cp Q-Y5F6_1M.se.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.bam
cp Q-Y5F6_1M.pe.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.pe.bam cp Q-Y5F6_1M.pe.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.pe.bam
cp Q-Y5F6_1M.se.sorted.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.bam cp Q-Y5F6_1M.se.sorted.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.bam
...@@ -60,18 +62,17 @@ cp Q-Y5F6_1M.pe.sorted.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.pe.sorted.bam ...@@ -60,18 +62,17 @@ cp Q-Y5F6_1M.pe.sorted.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.pe.sorted.bam
cp Q-Y5F6_1M.se.alignSummary.txt ./NEW_test_data/meta/Q-Y5F6_1M.se.alignSummary.txt cp Q-Y5F6_1M.se.alignSummary.txt ./NEW_test_data/meta/Q-Y5F6_1M.se.alignSummary.txt
cp Q-Y5F6_1M.pe.alignSummary.txt ./NEW_test_data/meta/Q-Y5F6_1M.pe.alignSummary.txt cp Q-Y5F6_1M.pe.alignSummary.txt ./NEW_test_data/meta/Q-Y5F6_1M.pe.alignSummary.txt
singularity run 'docker://bicf/gudmaprbkdedup:2.0.0' java -jar /picard/build/libs/picard.jar MarkDuplicates I=./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.bam O=Q-Y5F6_1M.se.deduped.bam M=Q-Y5F6_1M.se.deduped.Metrics.txt REMOVE_DUPLICATES=true singularity run 'docker://gudmaprbk/picard2.23.9:1.0.0' java -jar /picard/build/libs/picard.jar MarkDuplicates I=./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.bam O=Q-Y5F6_1M.se.deduped.bam M=Q-Y5F6_1M.se.deduped.Metrics.txt REMOVE_DUPLICATES=true
singularity run 'docker://bicf/gudmaprbkdedup:2.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.se.sorted.deduped.bam Q-Y5F6_1M.se.deduped.bam singularity run 'docker://gudmaprbk/picard2.23.9:1.0.0' samtools sort -@ 20 -O BAM -o Q-Y5F6_1M.se.sorted.deduped.bam Q-Y5F6_1M.se.deduped.bam
singularity run 'docker://bicf/gudmaprbkdedup:2.0.0' samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.deduped.bam Q-Y5F6_1M.se.sorted.deduped.bam.bai singularity run 'docker://gudmaprbk/picard2.23.9:1.0.0' samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.deduped.bam Q-Y5F6_1M.se.sorted.deduped.bam.bai
cp Q-Y5F6_1M.se.deduped.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.deduped.bam cp Q-Y5F6_1M.se.deduped.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.deduped.bam
cp Q-Y5F6_1M.se.sorted.deduped.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam cp Q-Y5F6_1M.se.sorted.deduped.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam
cp Q-Y5F6_1M.se.sorted.deduped.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam.bai cp Q-Y5F6_1M.se.sorted.deduped.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam.bai
cp Q-Y5F6_1M.se.deduped.Metrics.txt /NEW_test_data/meta/Q-Y5F6_1M.se.deduped.Metrics.txt
cp Q-Y5F6_1M.se.deduped.Metrics.txt ./NEW_test_data/meta/Q-Y5F6_1M.se.deduped.Metrics.txt cp Q-Y5F6_1M.se.deduped.Metrics.txt ./NEW_test_data/meta/Q-Y5F6_1M.se.deduped.Metrics.txt
for i in {"chr8","chr4","chrY"}; do for i in {"chr8","chr4","chrY"}; do
echo "samtools view -b ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam ${i} > Q-Y5F6_1M.se.sorted.deduped.${i}.bam; samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.deduped.${i}.bam Q-Y5F6_1M.se.sorted.deduped.${i}.bam.bai;"; echo "samtools view -b ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam ${i} > Q-Y5F6_1M.se.sorted.deduped.${i}.bam; samtools index -@ 20 -b Q-Y5F6_1M.se.sorted.deduped.${i}.bam Q-Y5F6_1M.se.sorted.deduped.${i}.bam.bai;";
done | singularity run 'docker://bicf/gudmaprbkdedup:2.0.0' parallel -j 20 -k done | singularity run 'docker://gudmaprbk/picard2.23.9:1.0.0' parallel -j 20 -k
cp Q-Y5F6_1M.se.sorted.deduped.chr4.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.chr4.bam cp Q-Y5F6_1M.se.sorted.deduped.chr4.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.chr4.bam
cp Q-Y5F6_1M.se.sorted.deduped.chr4.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.chr4.bam.bai cp Q-Y5F6_1M.se.sorted.deduped.chr4.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.chr4.bam.bai
cp Q-Y5F6_1M.se.sorted.deduped.chr8.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.chr8.bam cp Q-Y5F6_1M.se.sorted.deduped.chr8.bam ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.chr8.bam
...@@ -81,28 +82,30 @@ cp Q-Y5F6_1M.se.sorted.deduped.chrY.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M. ...@@ -81,28 +82,30 @@ cp Q-Y5F6_1M.se.sorted.deduped.chrY.bam.bai ./NEW_test_data/bam/small/Q-Y5F6_1M.
mkdir -p ./NEW_test_data/counts mkdir -p ./NEW_test_data/counts
mkdir -p ./NEW_test_data/counts/small mkdir -p ./NEW_test_data/counts/small
ln -s /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/geneID.tsv ln -s /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/geneID.tsv
ln -s /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/Entrez.tsv ln -s /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/Entrez.tsv
singularity run 'docker://bicf/subread2:2.0.0' featureCounts -T 20 -a /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/genome.gtf -G /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/genome.fna -g 'gene_name' --extraAttributes 'gene_id' -o Q-Y5F6_1M.se.countData -s 1 -R SAM --primary --ignoreDup ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam singularity run 'docker://gudmaprbk/subread2.0.1:1.0.0' featureCounts -T 20 -a /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/genome.gtf -G /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/genome.fna -g 'gene_name' --extraAttributes 'gene_id' -o Q-Y5F6_1M.se_countData -s 1 -R SAM --primary --ignoreDup ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam
singularity run 'docker://bicf/subread2:2.0.0' Rscript ./workflow/scripts/calculateTPM.R --count Q-Y5F6_1M.se.countData singularity run 'docker://gudmaprbk/subread2.0.1:1.0.0' Rscript ../workflow/scripts/calculateTPM.R --count Q-Y5F6_1M.se_countData
singularity run 'docker://bicf/subread2:2.0.0' Rscript ./workflow/scripts/convertGeneSymbols.R --repRID Q-Y5F6_1M.se singularity run 'docker://gudmaprbk/subread2.0.1:1.0.0' Rscript ../workflow/scripts/convertGeneSymbols.R --repRID Q-Y5F6_1M.se
cp Q-Y5F6_1M.se.featureCounts ./NEW_test_data/counts/small/Q-Y5F6_1M.se.countData cp Q-Y5F6_1M.se_countData ./NEW_test_data/counts/small/Q-Y5F6_1M.se_countData
cp Q-Y5F6_1M.se.countTable.csv ./NEW_test_data/counts/small/Q-Y5F6_1M.se.countTable.csv cp Q-Y5F6_1M.se.countTable.csv ./NEW_test_data/counts/small/Q-Y5F6_1M.se.countTable.csv
cp Q-Y5F6_1M.se.countTable.csv ./NEW_test_data/counts/small/Q-Y5F6_1M.se.tpmTable.csv cp Q-Y5F6_1M.se_tpmTable.csv ./NEW_test_data/counts/small/Q-Y5F6_1M.se_tpmTable.csv
mkdir -p ./NEW_test_data/bw mkdir -p ./NEW_test_data/bw
mkdir -p ./NEW_test_data/bw/small mkdir -p ./NEW_test_data/bw/small
singularity run 'docker://bicf/deeptools3.3:2.0.0' bamCoverage -p 20 -b ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam -o Q-Y5F6_1M.se.bw singularity run 'docker://gudmaprbk/deeptools3.5.0:1.0.0' bamCoverage -p 20 -b ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.bam -o Q-Y5F6_1M.se.bw
cp Q-Y5F6_1M.se.bw ./NEW_test_data/bw/small/Q-Y5F6_1M.se.bw cp Q-Y5F6_1M.se.bw ./NEW_test_data/bw/small/Q-Y5F6_1M.se.bw
mkdir -p ./NEW_test_data/fastqc mkdir -p ./NEW_test_data/fastqc
mkdir -p ./NEW_test_data/fastqc/small mkdir -p ./NEW_test_data/fastqc/small
singularity run 'docker://bicf/fastqc:2.0.0' ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz -o . singularity run 'docker://gudmaprbk/fastqc0.11.9:1.0.0' fastqc ./NEW_test_data/fastq/small/Q-Y5F6_1M.R1.fastq.gz -o .
cp Q-Y5F6_1M.R1_fastqc.html ./NEW_test_data/fastqc/small/Q-Y5F6_1M.R1_fastqc.html cp Q-Y5F6_1M.R1_fastqc.html ./NEW_test_data/fastqc/small/Q-Y5F6_1M.R1_fastqc.html
cp Q-Y5F6_1M.R1_fastqc.zip ./NEW_test_data/fastqc/small/Q-Y5F6_1M.R1_fastqc.zip cp Q-Y5F6_1M.R1_fastqc.zip ./NEW_test_data/fastqc/small/Q-Y5F6_1M.R1_fastqc.zip
echo -e "geneID\tchrom\ttx_start\ttx_end\tTIN" > Q-Y5F6_1M.se.sorted.deduped.tin.xls echo -e "geneID\tchrom\ttx_start\ttx_end\tTIN" > Q-Y5F6_1M.se.sorted.deduped.tin.xls
for i in {"chr8","chr4","chrY"}; do for i in {"chr8","chr4","chrY"}; do
echo "tin.py -i ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.${i}.bam -r /project/BICF/BICF_Core/shared/gudmap/references/GRCh38.p12.v31/bed/genome.bed; cat Q-Y5F6_1M.se.sorted.deduped.${i}.tin.xls | tr -s \"\\w\" \"\\t\" | grep -P \"\\t${i}\\t\";"; done | singularity run 'docker://bicf/rseqc3.0:2.0.0' parallel -j 20 -k >> Q-Y5F6_1M.se.sorted.deduped.tin.xls echo "tin.py -i ./NEW_test_data/bam/small/Q-Y5F6_1M.se.sorted.deduped.${i}.bam -r /project/BICF/BICF_Core/shared/gudmap/references/GRCm38.p6.vM22/bed/genome.bed; cat Q-Y5F6_1M.se.sorted.deduped.${i}.tin.xls | tr -s \"\\w\" \"\\t\" | grep -P \"\\t${i}\\t\";"; done | singularity run 'docker://gudmaprbk/rseqc4.0.0:1.0.0' parallel -j 20 -k >> Q-Y5F6_1M.se.sorted.deduped.tin.xls
cp Q-Y5F6_1M.se.sorted.deduped.tin.xls ./NEW_test_data/meta/Q-Y5F6_1M.se.sorted.deduped.tin.xls cp Q-Y5F6_1M.se.sorted.deduped.tin.xls ./NEW_test_data/meta/Q-Y5F6_1M.se.sorted.deduped.tin.xls
chgrp -R BICF_Core ./NEW_test_data
chmod -R 750 ./NEW_test_data
{
"bag": {
"bag_name": "Execution_Run_{rid}",
"bag_algorithms": [
"md5"
],
"bag_archiver": "zip",
"bag_metadata": {}
},
"catalog": {
"catalog_id": "2",
"query_processors": [
{
"processor": "csv",
"processor_params": {
"output_path": "Execution_Run",
"query_path": "/attribute/M:=RNASeq:Execution_Run/RID=17-BPAG/RID,Replicate_RID:=Replicate,Workflow_RID:=Workflow,Reference_Genone_RID:=Reference_Genome,Input_Bag_RID:=Input_Bag,Notes,Execution_Status,Execution_Status_Detail,RCT,RMT?limit=none"
}
},
{
"processor": "csv",
"processor_params": {
"output_path": "Workflow",
"query_path": "/entity/M:=RNASeq:Execution_Run/RID=17-BPAG/RNASeq:Workflow?limit=none"
}
},
{
"processor": "csv",
"processor_params": {
"output_path": "Reference_Genome",
"query_path": "/entity/M:=RNASeq:Execution_Run/RID=17-BPAG/RNASeq:Reference_Genome?limit=none"
}
},
{
"processor": "csv",
"processor_params": {
"output_path": "Input_Bag",
"query_path": "/entity/M:=RNASeq:Execution_Run/RID=17-BPAG/RNASeq:Input_Bag?limit=none"
}
},
{
"processor": "csv",
"processor_params": {
"output_path": "mRNA_QC",
"query_path": "/attribute/M:=RNASeq:Execution_Run/RID=17-BPAG/(RID)=(RNASeq:mRNA_QC:Execution_Run)/RID,Execution_Run_RID:=Execution_Run,Replicate_RID:=Replicate,Paired_End,Strandedness,Median_Read_Length,Raw_Count,Final_Count,Notes,RCT,RMT?limit=none"
}
},
{
"processor": "fetch",
"processor_params": {
"output_path": "assets/Study/{Study_RID}/Experiment/{Experiment_RID}/Replicate/{Replicate_RID}/Execution_Run/{Execution_Run_RID}/Output_Files",
"query_path": "/attribute/M:=RNASeq:Execution_Run/RID=17-BPAG/R:=RNASeq:Replicate/$M/(RID)=(RNASeq:Processed_File:Execution_Run)/url:=File_URL,length:=File_Bytes,filename:=File_Name,md5:=File_MD5,Execution_Run_RID:=M:RID,Study_RID:=R:Study_RID,Experiment_RID:=R:Experiment_RID,Replicate_RID:=R:RID?limit=none"
}
},
{
"processor": "fetch",
"processor_params": {
"output_path": "assets/Study/{Study_RID}/Experiment/{Experiment_RID}/Replicate/{Replicate_RID}/Execution_Run/{Execution_Run_RID}/Input_Bag",
"query_path": "/attribute/M:=RNASeq:Execution_Run/RID=17-BPAG/R:=RNASeq:Replicate/$M/RNASeq:Input_Bag/url:=File_URL,length:=File_Bytes,filename:=File_Name,md5:=File_MD5,Execution_Run_RID:=M:RID,Study_RID:=R:Study_RID,Experiment_RID:=R:Experiment_RID,Replicate_RID:=R:RID?limit=none"
}
}
]
}
}
\ No newline at end of file
{ {
"bag": { "bag": {
"bag_name": "Replicate_{rid}", "bag_name": "{rid}_inputBag",
"bag_algorithms": [ "bag_algorithms": [
"md5" "md5"
], ],
......
...@@ -84,7 +84,23 @@ process { ...@@ -84,7 +84,23 @@ process {
cpus = 2 cpus = 2
memory = '1 GB' memory = '1 GB'
} }
withName: outputBag { withName: uploadInputBag {
cpus = 1
memory = '1 GB'
}
withName: uploadExecutionRun {
cpus = 1
memory = '1 GB'
}
withName: uploadQC {
cpus = 1
memory = '1 GB'
}
withName: uploadProcessedFile {
cpus = 1
memory = '1 GB'
}
withName: uploadOutputBag {
cpus = 1 cpus = 1
memory = '1 GB' memory = '1 GB'
} }
......
...@@ -58,7 +58,19 @@ process { ...@@ -58,7 +58,19 @@ process {
withName: aggrQC { withName: aggrQC {
executor = 'local' executor = 'local'
} }
withName: outputBag { withName: uploadInputBag {
executor = 'local'
}
withName: uploadExecutionRun {
executor = 'local'
}
withName: uploadQC {
executor = 'local'
}
withName: uploadProcessedFile {
executor = 'local'
}
withName: uploadOutputBag {
executor = 'local' executor = 'local'
} }
} }
......
...@@ -74,10 +74,14 @@ custom_data: ...@@ -74,10 +74,14 @@ custom_data:
scale: false scale: false
format: '{}' format: '{}'
headers: headers:
Session Session:
Session ID description: ''
Pipeline Version Session ID:
Input description: 'Nextflow session ID'
Pipeline Version:
description: 'BICF pipeline version'
Input:
description: 'Input overrides'
rid: rid:
file_format: 'tsv' file_format: 'tsv'
section_name: 'RID' section_name: 'RID'
...@@ -88,10 +92,14 @@ custom_data: ...@@ -88,10 +92,14 @@ custom_data:
scale: false scale: false
format: '{}' format: '{}'
headers: headers:
Replicate Replicate:
Replicate RID description: ''
Experiment RID Replicate RID:
Study RID description: 'Replicate RID'
Experiment RID:
description: 'Experiment RID'
Study RID:
description: 'Study RID'
meta: meta:
file_format: 'tsv' file_format: 'tsv'
section_name: 'Metadata' section_name: 'Metadata'
...@@ -102,30 +110,43 @@ custom_data: ...@@ -102,30 +110,43 @@ custom_data:
scale: false scale: false
format: '{:,.0f}' format: '{:,.0f}'
headers: headers:
Source Source:
Species description: 'Metadata source'
Ends Species:
Stranded description: 'Species'
Spike-in Ends:
Raw Reads description: 'Single or paired end sequencing'
Assigned Reads Stranded:
Median Read Length description: 'Stranded (forward/reverse) or unstranded library prep'
Median TIN Spike-in:
Pipeline Version description: 'ERCC spike in'
Raw Reads:
description: 'Number of reads of the sequencer'
Assigned Reads:
description: 'Final reads after fintering'
Median Read Length:
description: 'Average read length'
Median TIN:
description: 'Average transcript integrity number'
ref: ref:
file_format: 'tsv' file_format: 'tsv'
section_name: 'Reference' section_name: 'Reference'
description: 'This is the referenec version information' description: 'This is the reference version information'
plot_type: 'table' plot_type: 'table'
pconfig: pconfig:
id: 'ref' id: 'ref'
scale: false scale: false
format: '{}' format: '{}'
headers: headers:
Species Species:
Genome Reference Consortium Build description: 'Reference species'
Genome Reference Consortium Patch Genome Reference Consortium Build:
GENCODE Annotation Release" description: 'Reference source build'
Genome Reference Consortium Patch:
description: 'Reference source patch version'
GENCODE Annotation Release:
description: 'Annotation release version'
tin: tin:
file_format: 'tsv' file_format: 'tsv'
section_name: 'TIN' section_name: 'TIN'
...@@ -135,16 +156,16 @@ custom_data: ...@@ -135,16 +156,16 @@ custom_data:
id: 'tin' id: 'tin'
headers: headers:
chrom chrom
0 - 9 1 - 10
10 - 19 11 - 20
20 - 29 21 - 30
30 - 39 31 - 40
40 - 49 41 - 50
50 - 59 51 - 60
60 - 69 61 - 70
70 - 79 71 - 80
80 - 89 81 - 90
90 - 99 91 - 100
sp: sp:
run: run:
...@@ -156,4 +177,4 @@ sp: ...@@ -156,4 +177,4 @@ sp:
ref: ref:
fn: 'reference.tsv' fn: 'reference.tsv'
tin: tin:
fn: '*.tin.hist.tsv' fn: '*_tin.hist.tsv'
...@@ -20,55 +20,67 @@ profiles { ...@@ -20,55 +20,67 @@ profiles {
process { process {
withName:getBag { withName:getBag {
container = 'bicf/gudmaprbkfilexfer:2.0.1_indev' container = 'gudmaprbk/deriva1.3:1.0.0'
} }
withName:getData { withName:getData {
container = 'bicf/gudmaprbkfilexfer:2.0.1_indev' container = 'gudmaprbk/deriva1.3:1.0.0'
} }
withName: parseMetadata { withName: parseMetadata {
container = 'gudmaprbk/python3:1.0.0' container = 'gudmaprbk/python3:1.0.0'
} }
withName: trimData { withName: trimData {
container = 'bicf/trimgalore:1.1' container = 'gudmaprbk/trimgalore0.6.5:1.0.0'
} }
withName: getRefInfer { withName: getRefInfer {
container = 'gudmaprbk/deriva1.3:1.0.0' container = 'gudmaprbk/deriva1.3:1.0.0'
} }
withName: downsampleData { withName: downsampleData {
container = 'bicf/seqtk:2.0.1_indev' container = 'gudmaprbk/seqtk1.3:1.0.0'
} }
withName: alignSampleData { withName: alignSampleData {
container = 'bicf/gudmaprbkaligner:2.0.1_indev' container = 'gudmaprbk/hisat2.2.1:1.0.0'
} }
withName: inferMetadata { withName: inferMetadata {
container = 'bicf/rseqc3.0:2.0.1_indev' container = 'gudmaprbk/rseqc4.0.0:1.0.0'
} }
withName: getRef { withName: getRef {
container = 'gudmaprbk/deriva1.3:1.0.0' container = 'gudmaprbk/deriva1.3:1.0.0'
} }
withName: alignData { withName: alignData {
container = 'bicf/gudmaprbkaligner:2.0.1_indev' container = 'gudmaprbk/hisat2.2.1:1.0.0'
} }
withName: dedupData { withName: dedupData {
container = 'bicf/gudmaprbkdedup:2.0.0' container = 'gudmaprbk/picard2.23.9:1.0.0'
} }
withName: countData { withName: countData {
container = 'bicf/subread2:2.0.0' container = 'gudmaprbk/subread2.0.1:1.0.0'
} }
withName: makeBigWig { withName: makeBigWig {
container = 'bicf/deeptools3.3:2.0.1_indev' container = 'gudmaprbk/deeptools3.5.0:1.0.0'
} }
withName: fastqc { withName: fastqc {
container = 'bicf/fastqc:2.0.1_indev' container = 'gudmaprbk/fastqc0.11.9:1.0.0'
} }
withName: dataQC { withName: dataQC {
container = 'bicf/rseqc3.0:2.0.1_indev' container = 'gudmaprbk/rseqc4.0.0:1.0.0'
} }
withName: aggrQC { withName: aggrQC {
container = 'bicf/multiqc1.8:2.0.1_indev' container = 'gudmaprbk/multiqc1.9:1.0.0'
}
withName:uploadInputBag {
container = 'gudmaprbk/deriva1.3:1.0.0'
}
withName:uploadExecutionRun {
container = 'gudmaprbk/deriva1.3:1.0.0'
} }
withName:outputBag { withName:uploadQC {
container = 'bicf/gudmaprbkfilexfer:2.0.1_indev' container = 'gudmaprbk/deriva1.3:1.0.0'
}
withName:uploadProcessedFile {
container = 'gudmaprbk/deriva1.3:1.0.0'
}
withName:uploadOutputBag {
container = 'gudmaprbk/deriva1.3:1.0.0'
} }
} }
......
This diff is collapsed.
File moved
...@@ -23,4 +23,4 @@ output <- merge(x=convert,y=countTable[,c("gene_name","gene_id","count","tpm")], ...@@ -23,4 +23,4 @@ output <- merge(x=convert,y=countTable[,c("gene_name","gene_id","count","tpm")],
colnames(output) <- c("GENCODE_Gene_Symbol","NCBI_GeneID","Ensembl_GeneID","count","tpm") colnames(output) <- c("GENCODE_Gene_Symbol","NCBI_GeneID","Ensembl_GeneID","count","tpm")
output <- output[,c(1,3,2,4:5)] output <- output[,c(1,3,2,4:5)]
write.table(output,file=paste0(opt$repRID,".tpmTable.csv"),sep=",",row.names=FALSE,quote=FALSE) write.table(output,file=paste0(opt$repRID,"_tpmTable.csv"),sep=",",row.names=FALSE,quote=FALSE)
import argparse
from deriva.core import ErmrestCatalog, get_credential, BaseCLI
import sys
import csv
def get_args():
parser = argparse.ArgumentParser()
parser.add_argument('-r', '--RID', help="replicate RID", required=True)
parser.add_argument('-t', '--table', help="source table", required=True)
parser.add_argument('-o', '--host', help="datahub host", required=True)
parser.add_argument('-c', '--cookie', help="cookie token", required=True)
args = parser.parse_args()
return args
def main(hostname, catalog_number, credential):
catalog = ErmrestCatalog('https', hostname, catalog_number, credential)
pb = catalog.getPathBuilder()
if args.table == 'mRNA_QC':
run_table = pb.RNASeq.mRNA_QC
elif args.table == "Processed_File":
run_table = pb.RNASeq.Processed_File
path = run_table.filter(run_table.RID == args.RID)
path.delete()
rid = args.RID
print(rid + " deleted")
if __name__ == '__main__':
args = get_args()
cli = BaseCLI("Custom RNASeq query", None, 1)
cli.remove_options(["--config-file"])
host = args.host
credentials = {"cookie": args.cookie}
main(host, 2, credentials)
\ No newline at end of file
...@@ -38,7 +38,7 @@ SOFTWARE_REGEX = { ...@@ -38,7 +38,7 @@ SOFTWARE_REGEX = {
'Trim Galore!': ['version_trimgalore.txt', r"version (\S+)"], 'Trim Galore!': ['version_trimgalore.txt', r"version (\S+)"],
'HISAT2': ['version_hisat2.txt', r"version (\S+)"], 'HISAT2': ['version_hisat2.txt', r"version (\S+)"],
'Samtools': ['version_samtools.txt', r"samtools (\S+)"], 'Samtools': ['version_samtools.txt', r"samtools (\S+)"],
'picard (MarkDuplicates)': ['version_markdups.txt', r"(\S\.\S{2}\.\S+)"], 'picard (MarkDuplicates)': ['version_markdups.txt', r"Version:(\S+)"],
'featureCounts': ['version_featurecounts.txt', r"featureCounts v(\S+)"], 'featureCounts': ['version_featurecounts.txt', r"featureCounts v(\S+)"],
'R': ['version_r.txt', r"R version (\S+)"], 'R': ['version_r.txt', r"R version (\S+)"],
'deepTools': ['version_deeptools.txt', r"deeptools (\S+)"], 'deepTools': ['version_deeptools.txt', r"deeptools (\S+)"],
......
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment