-
Notifications
You must be signed in to change notification settings - Fork 341
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Metrics of experiments with different tech implementation #34
Comments
@andronovhopf really great feedback! Most of it is under development already in the core-dvc:
The spreadsheet of experiments is another great idea. We should think about that. Re the spreadsheet... what would be your criteria to include an experiment into the table? How many of these would you expect to see here? |
@andronovhopf how and where do you specify the |
@shcheklein yes @dmpetrov, re: spreadsheet. Two ways of selecting experiments to display in a table come to mind:
Any other ideas? |
@andronovhopf did you run it like I like both the ways. If we do that:
|
Oh, I see that in the repo https://github.com/andronovhopf/cml_scratch |
@dmpetrov the pipeline has two stages ( |
Another observation: my project has two branches; on Now, because the hyperparameters I'm collecting are not the same as on Also, I know we are planning to do this eventually- but here's a case where being able to compare two commits on the same branch, instead of the head of two branches, would be great (as an additional option, not instead of). Since I want to test a few different numbers of neurons/epochs in the neural network. |
@andronovhopf nice observation. We had that discussion also. That every branch might be different implementations of the same problem to be solved. Like here a DNN vs Random forest. You can setup a different baseline and a baseline can be an specific commit sha. You can setup your baseline i.e to be HEAD~1 to compare your experiments with your previous one. And thats why the top five list came also in place to have a fast access to the same branch. In my personal experience, to solve your problem in your DNN branch change the baseline to master/dnn (supposing its called that way) and work with branches of that branch to adjust new parameters. |
Closed this is not relevant anymore. Belongs to the CML-DVC incarnation of CML |
This is a discussion point, not really an issue. I'm thinking about how metrics are displayed:
I definitely want to know that I'm comparing two experiments in which hyperparameters of my model (here, the maximum depth of a random forest classifier
max_depth
) changed. But, whereas it makes sense to have a "diff" presented for the accuracy metric, I'm not so sure it matters to have a diff present for the hyperparameters. It's not a number we're trying to optimize (unlike accuracy diffs) and visually, it makes the display more cluttered.I might suggest having a separate table for comparing hyperparameters that doesn't present diffs, just a side-by-side comparison. And then a table for comparing the output metrics, where I do care about the diff. Would this be challenging to implement? Maybe, for each distinct metric file, its own table? And then somewhere in project preferences a user could specify if we want diffs.
Another way of thinking about this is that if I had a spreadsheet of experiments I was trying to compare, I would lay it out this way:
And then perhaps highlight the row containing the best experiment (assuming that we can specify somehwere if we want + or - for the metric). If you want the diff explicitly calculated, maybe put it in its own field below the table.
The text was updated successfully, but these errors were encountered: