-
Notifications
You must be signed in to change notification settings - Fork 267
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Definition of scoring method unclear #4
Comments
Notwithstanding the unresolved conversation in #2, my preference is to use the F1 score as the true measure of performance, since the confusion matrix — while really nice and very practical in a real-world problem — opens another can of worms. Probably several cans, eg.
Definitely open to others' opinions on this. |
I thought the:
The easiest would be F1 score. Essentially a high F1 score would also mean that the well log is predicted correctly and "adjacency". Only going for high adjacency could mean that your well logs have all the "allowed adjacent" facies but a low F1 score i.e. wrong exact predictions. Curious to see what others run into here. |
I tested stratiefiedKfold. I remember that it gives much higher results Matt Hall wrote:
|
We now have STUART and CRAWFORD facies data so will score the contest based on F1 score against those known facies. Please see this notebook for a demo of the k-fold workflow in I am closing this issue now, because I think those were the main two concerns and — I think! — they have both been addressed. Please let me know if you disagree! |
Which score will be used to score an individual prediction?
accuracy(confusion_matrix)
accuracy_adjacent(confusion_matrix, adjacent_facies)
skm.metrics.f1_score(Y_true, Y_pred, sample="weighted")
accuracy*accuracy_adjacent?
The text was updated successfully, but these errors were encountered: