-
-
Notifications
You must be signed in to change notification settings - Fork 173
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Discussing correlation vs. causation with BlackBlaze data #275
Comments
I totally agree with you ( #187 (comment) ), see my quote from the BB data (and theres a link there). This is correlation not causation and I think to display it as "Failed" is kind of misleading. At best its a "warning" or a "red flag" that something might happen, not that it has. Thats why BB dont use most of these data points for their decisions about replacing drives, etc. |
Yeah, as @shamoon mentioned, there's been a lot of concern about how Backblaze data is used within Scrutiny. I'm working on some changes to the failure detection such that it'll be configurable in the UI, and you can selectively enable/disable the backblaze based failures and the thresholds. |
Just wanted to give everyone an update on the status of this issue. There's currently two tasks I'm working on:
The first task is partially complete. Here's what it currently looks like: The Smart status and Scrutiny status are differentiated in the expanding details panel. This is still a prototype. I think it works well, but I'd love to hear your thoughts. |
Is that a seagate drive ? Probably related to #255 then, and fixed in master to not show failed anymore. |
Took an incredibly long time, but as of On the dashboard settings panel, you can now change the "Device Status - Thresholds" between When changed to The description and UI for this functionality may be enhanced in the coming releases, but it is functional and working. Appreciate everyone's patience - this has been a long time coming. |
I just installed scrutiny and set it up for my 6 external HDDs. Some of them are more than 5 years old, some of them are brand new. I noticed that two of the old disks were marked as failed:
Disk 1:
Disk 2:
This looks pretty scary, given the high failure rate and its description:
However, it made me think. Does the failure rate mean that 20% of confirmed failed disks in BlackBlaze's dataset had this attribute at or worse than my value? And does this take into account how many of the healthy disks had this attribute with the same value? Because if we only look at the failed data, we're assuming that correlation is causation, which may be wrong. Ideally, I believe we'd want to report the difference between the healthy and failed disks instead. This may be what you're currently doing, but I have no clue, so please excuse any assumptions I made here.
Thanks a lot!
EDIT: Could you please share the source of the BlackBlaze data?
The text was updated successfully, but these errors were encountered: