-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
User narrative #28
Comments
dahak would take URLs for read files, but if they were present locally on the machine it would not use the URL or try to download them. Need to improve on that. Simply imposing the requirement that the user specify a URL would exclude users who want to use taco on local data - not sensible. So perhaps we could have an option to take a URL, or take an absolute path, and if the data is not available at the absolute path the workflow fails (i.e. if the user wants to provide their data locally, they must ensure their data is available on the local machine where the snakemake task will run). |
http://nih-data-commons.us/use-case-library/glossary/ Scientific objective: a description of a scientific process, told without reference to a specific technology solution. Focuses on resources required, method(s) followed, and outcome(s) targeted. User epic: a story, told from the user's perspective that captures a valuable set of interactions with tools (e.g. software solutions) in the service of achieving the scientific objective. User story: a specific piece of an epic that translates to a well-scoped and defined piece of software or data engineering work. |
Trimming ReadsScientific objective:
User epic: (already somewhat naturally broken up into user stories)
|
Relates to test-driven development. Develop a user narrative for taco, and develop command line interface around that.
For example: I want to trim some reads. How do I create/use taco-read-filtering? Can I use a local config/params file? Do I need the config/params file in the repo?
What about data - what if my reads live locally? What if they're available in scratch space on the cluster? (If we target AWS as a platform, that simplifies things somewhat, as we're always assuming a green-field deployment and thus assuming read data is going to be remote/in a bucket. But we're trying to keep taco platform agnostic.)
The text was updated successfully, but these errors were encountered: