You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Error executing process > 'hostremoval_input_fastq (BSH001.A0101.SG1)'
Caused by:
Process `hostremoval_input_fastq (BSH001.A0101.SG1)` terminated with an error exit status (1)
Command executed:
samtools index BSH001.A0101.SG1_PE.mapped.bam
extract_map_reads.py BSH001.A0101.SG1_PE.mapped.bam BSH001.A0101.SG1_R1_lanemerged.fq.gz -rev BSH001.A0101.SG1_R2_lanemerged.fq.gz -m remove -of BSH001.A0101.SG1_PE.mapped.hostremoved.fwd.fq.gz -or BSH001.A0101.SG1_PE.mapped.hostremove
d.rev.fq.gz -p 1
Command exit status:
1
Command output:
- Extracting mapped reads from BSH001.A0101.SG1_PE.mapped.bam
- Parsing forward fq file BSH001.A0101.SG1_R1_lanemerged.fq.gz
Command error:
Traceback (most recent call last):
File "/home/irina_marie_velsko/.nextflow/assets/nf-core/eager/bin/extract_map_reads.py", line 270, in <module>
File "/home/irina_marie_velsko/.nextflow/assets/nf-core/eager/bin/extract_map_reads.py", line 147, in parse_fq
File "/home/irina_marie_velsko/.nextflow/assets/nf-core/eager/bin/extract_map_reads.py", line 120, in get_fq_reads
File "/opt/conda/envs/nf-core-eager-2.3.5/lib/python3.7/site-packages/Bio/SeqIO/QualityIO.py", line 933, in FastqGeneralIterator
seq_string = handle_readline().rstrip()
File "/opt/conda/envs/nf-core-eager-2.3.5/lib/python3.7/site-packages/xopen/__init__.py", line 268, in readline
return self._file.readline(*args)
File "/opt/conda/envs/nf-core-eager-2.3.5/lib/python3.7/codecs.py", line 322, in decode
(result, consumed) = self._buffer_decode(data, self.errors, final)
MemoryError
Work dir:
/mnt/archgen/microbiome_calculus/abpCapture/03-preprocessing/set1_set3/work/e4/663badafbd377d9291bdb211a98525
Tip: when you have fixed the problem you can continue the execution adding the option `-resume` to the run command line
-[nf-core/eager] Pipeline completed with errors-
Expected behaviour
There should be enough memory for the host-mapped reads to be removed from the input files and the resulting host_removed fastq files to be written for both forward and reverse reads.
Log files
Have you provided the following extra information/files:
The command used to run the pipeline
The .nextflow.log file
The exact error: see above
System
Hardware: HPC
Executor: sge
OS: Linux
Version: Ubuntu 20.04.3 LTS
Nextflow Installation
Version: 20.10.0 build 5430
Container engine
Engine: Singularity
version:
Image tag: nfcore/eager:2.3.5
Additional context
I tried to increase the memory from 32GB by adjusting the lines
Note: these tools are probably indeed more robust, but we need to consider that it might be unwanted to just 'replace' reads or variants with ref. genome ones. If someone wanted to reanalyse e.g. calculus for human DNA, they may not realise they are lookin at 'fake' sequence.
This is sort of 'tampering' with the FASTQ file in a misleading sense. Therefore I would still rather have a NNN replacement or entire removal (will need to check if the tools support this).
Check Documentation
I have checked the following places for your error:
Description of the bug
hostremoval_input_fastq
fails on larger samples b/c it runs out of memorySteps to reproduce
Steps to reproduce the behaviour:
Expected behaviour
There should be enough memory for the host-mapped reads to be removed from the input files and the resulting host_removed fastq files to be written for both forward and reverse reads.
Log files
Have you provided the following extra information/files:
.nextflow.log
fileSystem
Nextflow Installation
Container engine
Additional context
I tried to increase the memory from 32GB by adjusting the lines
in the
.command.run
file. It ran with 180GB as written above, but theqacct
record saysmaxvmem 120.745GB
.The text was updated successfully, but these errors were encountered: