Skip to content

Set up AWS megatests #80

@ewels

Description

@ewels

AWS megatests is now running nicely and we’re trying to set up all (most) nf-core pipelines to run a big dataset. We need to identify a set of public data to run benchmarks for the pipeline.

The idea is that this will run automatically for every release of the nf-core/hic pipeline. The results will then be publicly accessible from s3 and viewable through the website: https://nf-co.re/hic/results - this means that people can manually compare differences in output between pipeline releases if they wish.

We need a dataset that is as “normal” as possible, mouse or human, sequenced relatively recently and with a bunch of replicates etc. It can be a fairly large project

I'm hoping that @nservant can help here, but suggestions from anyone and everyone are more than welcome! ✋🏻

In practical terms, once decided we need to:

  • Upload the FastQ files to s3: s3://nf-core-awsmegatests/hic/input_data/ (I can help with this)
  • Update test_full.config to work with these file paths
  • Check .github/workflows/awsfulltest.yml (should be no changes required I think?)
  • Merge, and try running the dev branch manually

Metadata

Metadata

Assignees

Labels

enhancementNew feature or request

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions