Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Refine the observation generation through internal testing #82

Open
hellais opened this issue Aug 23, 2024 · 3 comments
Open

Refine the observation generation through internal testing #82

hellais opened this issue Aug 23, 2024 · 3 comments

Comments

@hellais
Copy link
Member

hellais commented Aug 23, 2024

The goal of this issue is to collect initial feedback internally on how the observation generation and analysis is currently working.

There is a test instance of this running here:

For example if you want to see the observations and analysis generated for the recent nepal report measurements, you can do:

This issue is a place to collect all relevant bugs that are spotted and identify any issues we might find. The primary focus is on observation generation, but it's also useful to start cataloging any specific cases where the analysis is grossly off.

One important caveat (which mostly applies to analysis) is that the processing is done direct in real-time when you access the page. This means that if a ground truth database does not exist, the analysis might be a bit off compared to how it will look like in a real production scenario.

@sitinurliza95
Copy link

Checking through some measurements that I found with issues so far:

Seems ok. But will need to go through more HK measurements as they have the most issues

@sitinurliza95
Copy link

An OK measurement that is a blocking - blocked 0.6 (not sure why it's been showing a lot of OK on Explorer. I've just added the SG fingerprints ooni/blocking-fingerprints#17)
https://data.ooni.org/analysis/m/20240401153454.011394_SG_webconnectivity_74d14611a801555f

@hellais
Copy link
Member Author

hellais commented Oct 4, 2024

Thanks for documenting these cases! The ones marked as blocked 0.2 are instances in which the fingerprint is inconsistent with the country hence we mark it as 0.2 blocked and use the down indicator to flag it as "down" in the sense that the probe is badly located or misconfigured.

It might make sense to support these cases to have an extra parameter which is used specifically to flag bad measurements with also some likelyhood distribution. Similar to what we do ATM with the failed flag.

For the NXDOMAIN one, I will have to look at that more carefully.

I think the one marked as blocked 0.6, given we don't have the fingerprint in our DB, the value seems reasonable, though maybe it ought to be weighed a bit differently.

All in all thanks for highlighting these cases they are super helpful!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants