Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GetOrganelle aborts with error ("Killed") #264

Closed
3 tasks done
marcelglueck opened this issue May 9, 2023 · 3 comments
Closed
3 tasks done

GetOrganelle aborts with error ("Killed") #264

marcelglueck opened this issue May 9, 2023 · 3 comments

Comments

@marcelglueck
Copy link

First check

  • I used the GitHub search to find a similar issue and didn't find it.
  • I searched GetOrganelle.wiki context, especially the FAQ and browsed the examples to confirm it is unexpected to happen.
  • I have updated GetOrganelle to the latest version GitHub release

Describe the bug
In making read indexes, the program exits with "13766 Killed". No further information is given. I am wondering whether this might be a RAM issue.

The following command was executed:
get_organelle_from_reads.py -1 $read1 -2 $read2 -o $dout/$subdout -R 40 -k 21,45,65,85,105,135 -F $seedDB -t 14 --config-dir ${dbloc}${seedDB} --overwrite --max-reads INF --reduce-reads-for-coverage 800 -J 1 -M 1
All_species_plastids_assembly.o10946194-6.txt

Additional context
GetOrganelle is containerized (Apptainer) and runs as part of an array job. Each subjob is provided 14 cores and 30 GB of RAM. So far, other subjobs have not aborted with this error and are running just fine.

@marcelglueck
Copy link
Author

Update: Meanwhile, the other subjobs aborted as well with the same error message.

@JianjunJin
Copy link
Collaborator

JianjunJin commented May 9, 2023

Thanks for reaching out with the details. I agree it might be a RAM issue.

It is hard to know the appropriate RAM before a run, especially for large datasets. Given this dataset and the log, a slightly larger RAM (e.g. 40G) could help surmount the index-making process. Unlike plants, it usually uses much less memory for animal_mt in downstream steps, but again there is no guarantee of 40G for downstream steps and all your jobs. If memory is not a big issue in your cluster, I recommend increasing the RAM request. Otherwise, see here to save memory consumption.


PS: I've learned that there is quite room for RAM improvement without efficiency loss from the GetOrganelle development side. But I am not available to make a big release for this type of improvement recently.

@marcelglueck
Copy link
Author

Thank you for your fast response. Indeed, increasing RAM did the trick.
Analysing the logs, I discovered something odd. Kmer sizes specified were 21,45,65,85,105,135 (see command above). However, the logs tell a different story with "Setting '-k 21,45,65,85,105'", dropping the last kmer length. This also becomes apparent in the graph files containing K105 in their files names. Is 105 the longest kmer length supported? If I grasped it correctly, the likelihood of obtaining circular genomes can be increased by incorporating longer kmers. Hence, I wished to incorporate longer kmers in the assembly.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants