-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error running assembly workflow on Ubuntu: Error in rule assembly_megahit: #113
Comments
I have a question about the Assembly workflow in the Really Quick Copy-and-Paste Quick Start: I noticed that the input for the assembly is the same (raw data) that is called for the read filtering workflow. Shouldn't the input for assembly workflow be the output of the readfiltering workflow (i.e. filtered/trimmed reads rather than the raw reads)? The same raw data are also called for the inputs with the next workflows ( comparison, taxonomic classification)? Is this something that the user should be updating with each progressive workflow? Could this be a possible source of the error for that I am seeing above for the assembly? I appreciate any clarification you can provide! |
The input files are the same, but the trimming workflow is run on them prior
to assembly. (If trimming has already happened, it is not re-run.)
You can kind of infer this from the 'qual' argument in the assembly workflow
JSON, and also by watching what snakemake does, but we will make sure to
add this to the docs.
(Snakemake says:
Job 1: --- Assembling quality trimmed reads with Megahit)
|
In order to diagnose the error, we might need the file |
Thanks for the explanation! Here is the .log file that you requested above: MEGAHIT v1.1.2 |
cool, #115 fixes this! (I had the branch ready but had forgotten to make the PR!) |
Re-kicked this off this morning and seems to be running fine with no errors so far! :) |
As per assembly workflow in Really Quick Copy-And-Paste Quick Start I copied the json file specified and execute the following which gave me the following error:
$ export SINGULARITY_BINDPATH="data:/data"
$ snakemake -p --use-singularity --configfile=config/custom_assembly_workflow.json assembly_workflow_all
Error:
Building DAG of jobs...
Pulling singularity image docker://quay.io/biocontainers/spades:3.11.1--py27_zlib1.2.8_0.
Pulling singularity image docker://quay.io/biocontainers/megahit:1.1.2--py35_0.
Using shell: /bin/bash
Provided cores: 1
Rules claiming more threads will be scaled down.
Job counts:
count jobs
4 assembly_megahit
4 assembly_metaspades
1 assembly_workflow_all
9
Job 6: --- Assembling quality trimmed reads with Megahit
rm -rf data/SRR606249_subset25.trim2_megahit && megahit -t 1 --memory 0.20 -1 /data/SRR606249_subset25_1.trim2.fq.gz -2 /data/SRR606249_subset25_2.trim2.fq.gz --out-prefix=SRR606249_subset25.trim2_megahit -o /data/SRR606249_subset25.trim2_megahit && mv /data/SRR606249_subset25.trim2_megahit/SRR606249_subset25.trim2_megahit.contigs.fa /data/SRR606249_subset25.trim2_megahit.contigs.fa
Activating singularity image /home/user/dahak_2018/dahak/workflows/.snakemake/singularity/bfd669a63b585d366276296fdcd11501.simg
7.791Gb memory in total.
Using: 1.558Gb.
MEGAHIT v1.1.2
--- [Tue Jul 17 21:49:41 2018] Start assembly. Number of CPU threads 1 ---
--- [Tue Jul 17 21:49:41 2018] Available memory: 8365150208, used: 1673030041
--- [Tue Jul 17 21:49:41 2018] Converting reads to binaries ---
b' [read_lib_functions-inl.h : 209] Lib 0 (/data/SRR606249_subset25_1.trim2.fq.gz,/data/SRR606249_subset25_2.trim2.fq.gz): pe, 26715952 reads, 101 max length'
b' [utils.h : 126] Real: 71.9618\tuser: 23.4359\tsys: 10.3257\tmaxrss: 155080'
--- [Tue Jul 17 21:50:53 2018] k-max reset to: 119 ---
--- [Tue Jul 17 21:50:53 2018] k list: 21,29,39,59,79,99,119 ---
--- [Tue Jul 17 21:50:53 2018] Extracting solid (k+1)-mers for k = 21 ---
--- [Tue Jul 17 22:13:00 2018] Building graph for k = 21 ---
Error occurs when running "builder build" for k = 21; please refer to /data/SRR606249_subset25.trim2_megahit/SRR606249_subset25.trim2_megahit.log for detail
[Exit code 1]
Error in rule assembly_megahit:
jobid: 6
output: data/SRR606249_subset25.trim2_megahit.contigs.fa
log: data/SRR606249_subset25.trim2_megahit.log
RuleException:
CalledProcessError in line 148 of /home/user/dahak_2018/dahak/workflows/assembly/Snakefile:
Command 'singularity exec --home /home/user/dahak_2018/dahak/workflows /home/user/dahak_2018/dahak/workflows/.snakemake/singularity/bfd669a63b585d366276296fdcd11501.simg bash -c ' set -euo pipefail; rm -rf data/SRR606249_subset25.trim2_megahit && megahit -t 1 --memory 0.20 -1 /data/SRR606249_subset25_1.trim2.fq.gz -2 /data/SRR606249_subset25_2.trim2.fq.gz --out-prefix=SRR606249_subset25.trim2_megahit -o /data/SRR606249_subset25.trim2_megahit && mv /data/SRR606249_subset25.trim2_megahit/SRR606249_subset25.trim2_megahit.contigs.fa /data/SRR606249_subset25.trim2_megahit.contigs.fa '' returned non-zero exit status 1.
File "/home/user/dahak_2018/dahak/workflows/assembly/Snakefile", line 148, in __rule_assembly_megahit
File "/home/user/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 56, in run
Shutting down, this might take some time.
Exiting because a job execution failed. Look above for error message
Complete log: /home/user/dahak_2018/dahak/workflows/.snakemake/log/2018-07-17T214845.783217.snakemake.log
The text was updated successfully, but these errors were encountered: