-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question #8
Comments
Hi, can you clone the repo again and try to install it again? I have just updated some dependencies |
I tried, but it was not working so I installed all the dependencies manually one by one. However, it does not allow me to install snippy and centrifuge. I am trying these on Anaconda. |
It shows the following error after running a while. Touching output file fastq/SRR1210481.snippy. [Wed Jul 27 14:59:49 2022] /usr/bin/bash: CENTRIFUGE_DEFAULT_DB: unbound variable Shutting down, this might take some time. |
I reinstalled snippy nut it is still showing following error. /usr/bin/bash: CENTRIFUGE_DEFAULT_DB: unbound variable Shutting down, this might take some time. |
The error message is with Centrifuge, not snippy. Before running the pipeline, you need to download the centrifuge database, then set it up as shown in the README.md file and also set up Krona taxonomy. Can you confirm that you have completed these steps? |
I performed all these steps, the centrifuge database is installed wget -c ftp://ftp.ccb.jhu.edu/pub/infphilo/centrifuge/data/p_compressed+h+v.tar.gz |
I also setup Krona with the following steps: rm -rf ~/anaconda3/envs/bagep/opt/krona/taxonomy snakemake --config ref=enterococcus_genome.fasta However, it is showing an error: Error in rule snippy: Removing output files of failed job snippy since they might be corrupted: |
What version of snippy do you have installed and can you share the snippy
log file?
Idowu Olawoye
about.me/idowu
…On Thu, 28 Jul 2022, 17:28 Manoj Kumar, ***@***.***> wrote:
I also setup Krona with the following steps:
rm -rf ~/anaconda3/envs/bagep/opt/krona/taxonomy
mkdir -p ~/krona/taxonomy
ln -s ~/krona/taxonomy/ ~/miniconda3/envs/bagep/opt/krona/taxonomy
ktUpdateTaxonomy.sh ~/krona/taxonomy
snakemake --config ref=enterococcus_genome.fasta
*However, it is showing an error:*
Error in rule snippy:
jobid: 755
output: fastq/SRR1210481/, fastq/SRR1210481.snippy
shell:
snippy --force --cleanup --outdir fastq/SRR1210481/ --ref
enterococcus_genome.fasta --R1 fastp/fastq/SRR1210481_R1.fastq.gz.fastp
--R2 fastp/fastq/SRR1210481_R2.fastq.gz.fastp
(exited with non-zero exit code)
Removing output files of failed job snippy since they might be corrupted:
fastq/SRR1210481/
Shutting down, this might take some time.
Exiting because a job execution failed. Look above for error message
Complete log:
/media/mmk6053/Data/Manoj_data/Entrococcus_project/BAGEP/.snakemake/log/2022-07-28T122529.877538.snakemake.log
—
Reply to this email directly, view it on GitHub
<#8 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AJNJOXJ33OE6N3BBWEQK5WLVWKYLTANCNFSM54MHSNWQ>
.
You are receiving this because you were assigned.Message ID:
***@***.***>
|
I am using snippy 4.4.3. Where I can find a snippy log file since the pipeline isn't started the snippy? |
Did other steps run without issues? Also is your reference genome in the directory level where the snakefile is in? |
Pipeline starts with generating a message "Filtering fastQ files by trimming low quality reads using fastp". It generates a folder "fastp" and two R1 and R2 files, after it stops. |
Here is the complete run. (bagep) mmk53@A8-VT-MMK53-U1:/media/Data/Manoj_data/Entrococcus_project/BAGEP$ snakemake --config ref=enterococcus_genome.fasta [Thu Jul 28 14:33:11 2022] Read1 before filtering: Read2 before filtering: Read1 after filtering: Read2 after filtering: Filtering result: Duplication rate: 0.754108% Insert size peak (evaluated by paired-end reads): 144 JSON report: fastp.json fastp -i fastq/SRR1210481_R1.fastq.gz -I fastq/SRR1210481_R2.fastq.gz -o fastp/fastq/SRR1210481_R1.fastq.gz.fastp -O fastp/fastq/SRR1210481_R2.fastq.gz.fastp [Thu Jul 28 14:34:05 2022] /usr/bin/bash: CENTRIFUGE_DEFAULT_DB: unbound variable Shutting down, this might take some time. |
It seems, it is showing some error in filtering as low quality reads. However, these data I used separately with Snippy and it ran successfully. |
It appears you are running the workflow with 1 core. You can split the job across multiple threads depending on how many you have available. Try: Also the log message shows that If a stage in the pipeline fails, it will truncate the entire process |
Hi, |
Hi, Sorry for slow response. I will come back to my analysis at BAGEP. I got stuck in some other tasks. |
Hi, when I run the pipeline with --core 40, it occupies all my system's memory (~124G). Eventually, the system becomes stop. snakemake --cores 40 --config ref=enterococcus_genome.fasta |
Hi,
I am guessing that's because of the huge number of samples you're running
and the intermediary files that were generated. You might want to free up
some space or run it on an external drive with extra room. The pipeline
would clean up majority of the files at the end of the analysis and you can
also delete the files generated by fastp at the end of the run as well
Idowu Olawoye
about.me/idowu
…On Fri, 5 Aug 2022, 17:06 Manoj Kumar, ***@***.***> wrote:
Hi, when I run the pipeline with --core 40, it occupies all my system's
memory (~124G). Eventually, the system becomes stop.
snakemake --cores 40 --config ref=enterococcus_genome.fasta
—
Reply to this email directly, view it on GitHub
<#8 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AJNJOXPI27KJ3MJ6HKPTCZTVXU3YDANCNFSM54MHSNWQ>
.
You are receiving this because you were assigned.Message ID:
***@***.***>
|
Hi, Filtering result: fastp -i fastq/ERR4230412_R1.fastq.gz -I fastq/ERR4230412_R2.fastq.gz -o fastp/fastq/ERR4230412_R1.fastq.gz.fastp -O fastp/fastq/ERR4230412_R2.fastq.gz.fastp |
What does the log file look like? The fastp completed successfully but a rule of thumb is to pinpoint why the pipeline failed |
I could not find the log file. I suspect the pipeline was installed properly. Let me see the installing process again.. |
Hi,
This error is because you haven't set up your centrifuge database to that
variable. Have you downloaded the database to your computer?
Idowu Olawoye
about.me/idowu
…On Wed, 27 Jul 2022, 20:01 Manoj Kumar, ***@***.***> wrote:
It shows the following error after running a while.
Touching output file fastq/SRR1210481.snippy.
[Wed Jul 27 14:59:49 2022]
Finished job 755.
2 of 1221 steps (0.16%) done
[Wed Jul 27 14:59:49 2022]
Job 512: Taxonomic classification of processed reads using centrifuge
/usr/bin/bash: CENTRIFUGE_DEFAULT_DB: unbound variable
[Wed Jul 27 14:59:49 2022]
Error in rule centrifuge:
jobid: 512
output: taxonomy/fastq/SRR1210481-report.txt,
taxonomy/fastq/SRR1210481-result.txt
shell:
centrifuge -p 4 -x $CENTRIFUGE_DEFAULT_DB -1
fastp/fastq/SRR1210481_R1.fastq.gz.fastp -2
fastp/fastq/SRR1210481_R2.fastq.gz.fastp --report-file
taxonomy/fastq/SRR1210481-report.txt -S taxonomy/fastq/SRR1210481-result.txt
(exited with non-zero exit code)
Shutting down, this might take some time.
Exiting because a job execution failed. Look above for error message
Complete log:
/media/mmk6053/Data/Manoj_data/Entrococcus_project/BAGEP/.snakemake/log/2022-07-27T145131.709376.snakemake.log
—
Reply to this email directly, view it on GitHub
<#8 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AJNJOXM2ZMDTDMJ3LOI33EDVWGBQFANCNFSM54MHSNWQ>
.
You are receiving this because you were assigned.Message ID:
***@***.***>
|
I am trying to install BAGEP with the following command but it is running at "solving environment" for very long.
conda env create -f environment.yml
The text was updated successfully, but these errors were encountered: