-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Closed
Labels
feature requestRequesting a new featureRequesting a new featurep2-mediumMedium priority, should be done, but less importantMedium priority, should be done, but less important
Description
Hello,
my current ML project runs a DVC pipeline which trains a model for several different segments of the data, and produces a metrics file for each one.
I want to be able to track how these metrics evolve over time; however, the segments of data that are evaluated in each run depend on some parameters on the first step of the pipeline, and so the number of output metrics files changes every time I run the pipeline.
I have tried several approaches:
- adding
-m metricsto my DVC run command, where metrics/ is a folder where I store all my metrics files. This worked, but when callingdvc metrics show -a -Rit simply showed the DVC file for that folder, instead of parsing the individual metrics. - Using a wildcard -m metrics/* - this simply doesn't work.
- Setting the metrics directory as an output with
-o, and then adding each file as a metric using a script loop: I get the following errorERROR: failed to add metric file 'metrics/metric-1.json' - unable to find DVC-file with output 'metrics/metric-1.json
I am out of ideas! Is this functionality supported, or do I need to have a fixed number of metrics files for each pipeline?
efiop, dmpetrov and shcheklein
Metadata
Metadata
Assignees
Labels
feature requestRequesting a new featureRequesting a new featurep2-mediumMedium priority, should be done, but less importantMedium priority, should be done, but less important