-
Notifications
You must be signed in to change notification settings - Fork 910
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Evaluating experiment tracking adoption #1390
Comments
Initial report up from @NeroOkwa, I'll be making some edits to it. |
Experiment Tracking Research - ResultsGoalThe goal of this research task is to understand why Kedro Experiment Tracking is not being adopted Research QuestionWhy isn't Kedro ET being used, according to top most reasons? SummaryThe reason for the low adoption of Kedro Experiment Tracking is because users prefer other tools that have specific features that solve their 'job to be done'. Below are the top most reasons (based on frequency) that if addressed would increase user adoption of Kedro Experiment Tracking. The top 3 reasons are:
The other reasons (and supporting quotes) are shown in the results section below.Note: A demo session was also conducted for 12 users who had previously not used kedro experiment tracking. Output ResultsReasons sorted according to frequency of mention: Reason 1 - Ability to save and link images of plots/model artefacts to an experiment. This would provide users with more insight (images and metrics together) to track/compare the evolution of runs across a timeline - 5 users
Reason 2 - Visualisation: Ability to show plot /comparison graphs/hyper parameters to evaluate metrics tradeoff - 4 users
Reason 3 - Having server/ other storage database would enable multi-user collaboration on an experiment, for a user and their team - 4 users + 1 slack user
Reason 4 - Having kedro - viz as a standalone product would enable users to integrate kedro-viz with other tools (such as MLflow suite) for non-kedro projects - 2 users
Reason 5 - Improved loading and the ability to loop over parameters. This eliminates the need to re-enter the parameters, and also enables the user run different experiments. Both would simplify the user's workflow - 2 users
Reason 6 - Ability to visualise large pipelines for inspection and presentation by CSTs - 1 user
Reason 7 - Ability to parse a name to an experiment would enable the output comparisons of different configs- 1 user
Reason 8 - Division between metrics and parameters for tracked datasets (like in MLflow) makes comparing runs easier. - 1 user
Reason 9 - Ability to delete experiment tracking entries that are not useful - 1 user + 1 slack user
Reason 10 - Using older versions of Kedro/dependencies - 1 user
Reason 11 - Ability to address feature drift/input drift would reduce the current high frequency of client reporting- 1 user
|
I'll close this 🥳 This research is complete. |
Description
We shipped the first iteration of experiment tracking towards the end of November and we would like to understand why the feature has not been adopted. @NeroOkwa is leading this fantastic work.
Hypotheses
We suspect that Experiment Tracking is not being used because:
Possible Implementation
We're going to structure our research by finding target populations:
kedro viz
from their CLIWe're going to approach these user groups in the following ways:
kedro viz
from their CLI, we can schedule interviews and potentially run a survey with this groupActions to do:
kedro-viz
users)— quantity vs quality based on user groups (surveys vs 1to1s)kedro-viz
usersMeasure of Success in 3 Months
Additional data
From @yetudada to @Galileo-Galilei: You were also part of our user testing for Experiment Tracking in Kedro. I want to show you how to set it up (unfortunately it’s only available from Kedro 0.17.5) and the demo.
From @Galileo-Galilei to @yetudada: We have not been using it for now for several reasons:
kedro==0.16.5
for now because we have legacy projects in production and we do not want to add extra maintenance burden (we have dozens of projects in production + internal plugins , so migration is not a cost we can pay very often). We plan to move to 0.18.x by the end of the year, but we want to wait a little to assess the migration impacts and be sure the version is stable.mlflow
in production, and we don’t have time for testing extensively this new functionality for now. I think we will not use it before 2023.The text was updated successfully, but these errors were encountered: