Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

create a table with all tool settings used for the paper? #37

Open
alicemchardy opened this issue Jul 30, 2016 · 7 comments
Open

create a table with all tool settings used for the paper? #37

alicemchardy opened this issue Jul 30, 2016 · 7 comments
Assignees

Comments

@alicemchardy
Copy link
Contributor

We do have all tools fully reproducible in docker containers, however - for the making it easier for the reader, we could include a table with all tool parameter settings based on these dockers for the supplement. Any thoughts on this?

@sjanssen2
Copy link
Contributor

If we go into this direction, we should also include used inputs like databases. Furthermore, it is not always clear what type of input sequences could be used for some of the profiling tools. Besides raw reads some offer the option to input contigs (I can't see how one would end up with the same abundances as with reads) but it is a situation where we could guide potential uses.

@sjanssen2
Copy link
Contributor

I do know which parameters I used for those containers I compiled, but I am not sure if it is strait forward to investigate the parameter settings in partitcipant submitted containers - if we have those at all.
In what depth do we want to explain the underlying algorithms of the tools in our competition? If we don't, it might be hard to grasp what the parameter settings are good for.

@sjanssen2
Copy link
Contributor

to make results comparable, I often had to add an additional conversion step to the actual profiling tool to e.g. make it speak NCBI tax IDs instead of GreenGenes strings. Thus, the containers are not pure wrapper to the tools. We should clarify that to the users.

@alicemchardy
Copy link
Contributor Author

Good points. How about a table with the following columns

tool with version and reference / parameters used (where deviating from default) /modifications included by us / reference databases used (link to our collection)/ keywords to its approach

@p-hofmann
Copy link
Contributor

Some tools are part of a pipeline, using blast, diamond, bowtie2, kallisto as a preparation step.
Either we need to include a column for like 'additional info' or we could see it as part of the parameter column?

@sjanssen2
Copy link
Contributor

I started to compile this table: https://docs.google.com/spreadsheets/d/12flNC5JlhVl-qY_O9CSUs6Zauj6HWNU5NlC3kUn7WQc/edit?usp=sharing
It is hard to find out settings for the submitted dockerized tools. Do you think we could reach out to the submitters and ask them to fill out the missing information?

@alicemchardy
Copy link
Contributor Author

Good idea. I made a copy of the document the CAMI evaluation workshop folder and called it now CAMI_tool_settings - once the draft is circulated, developers should add this

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants