-
Notifications
You must be signed in to change notification settings - Fork 60
Description
AWS megatests is now running nicely and we’re trying to set up all (most) nf-core pipelines to run a big dataset. We need to identify a set of public data to run benchmarks for the pipeline.
The idea is that this will run automatically for every release of the nf-core/hic pipeline. The results will then be publicly accessible from s3 and viewable through the website: https://nf-co.re/hic/results - this means that people can manually compare differences in output between pipeline releases if they wish.
We need a dataset that is as “normal” as possible, mouse or human, sequenced relatively recently and with a bunch of replicates etc. It can be a fairly large project
I'm hoping that @nservant can help here, but suggestions from anyone and everyone are more than welcome! ✋🏻
In practical terms, once decided we need to:
- Upload the FastQ files to s3:
s3://nf-core-awsmegatests/hic/input_data/
(I can help with this) - Update
test_full.config
to work with these file paths - Check
.github/workflows/awsfulltest.yml
(should be no changes required I think?) - Merge, and try running the
dev
branch manually