-
Notifications
You must be signed in to change notification settings - Fork 551
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Active Learning sampling quality #983
Comments
can you try the better sampling branch and let me know if that is giving you better results? |
Hi again,
|
Looks like you are doing record linkage and not deduping, i haven't updated the code for that code path yet. |
Yes, exactly. I'll wait then, thank you. |
closing for now due to lack of feedback. |
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
After using the dedupe library for a while in the context of video content reconciliation, we encountered some situations where the Active Learning sampling is very poor. This makes it difficult to built a good training set for the classifier and by consequence the reconciliation results are poor as well.
For instance, we made some tests and tried to reconcile 2 well known public data provider (iMDB and TMDB), which contain reciprocal references to be used as ground truth and good metadata, plus we could build the dataset knowing that all entries could be reconciled in a
many-to-one
approach (set 1 is contained in set 2 -> 100% recall is theoretically possible).We tried to reconcile episodes and used few fields in the process (episode title, series title, season number, episode number, series year). The Active Learning sampling were quite balanced between positive and negative examples, therefore it has been quite effortless to collect 10 samples of positive and negative pairs. The final results were quite satisfying as well: recall 78%, precision 98%. Moreover, by scrolling through the results, we noticed that the model learned to ignore the episode title field, which were not consistent between datasets.
Afterwards we decided to perform a second test by removing the episode title field, but keeping everything else as in the previous test (same dataset, same configurations). This time the Active Learning sample were quite poor: almost all pairs were wrong (it took more than 200 pairs to obtain 8 positive). The final reconciliation in this case were also poor: recall 15% and precision 91%.
I would ask then if it is possible to mitigate this kind of issues:
Thank you for your great work,
Antonio
The text was updated successfully, but these errors were encountered: