Skip to content
This repository has been archived by the owner on May 16, 2023. It is now read-only.

Follow up on Exposure Notification API testing by Fraunhofer IIS #394

Open
kbobrowski opened this issue Aug 1, 2020 · 12 comments
Open

Follow up on Exposure Notification API testing by Fraunhofer IIS #394

kbobrowski opened this issue Aug 1, 2020 · 12 comments
Labels
question Further information is requested

Comments

@kbobrowski
Copy link

Your Question

Thanks for uploading API testing documentation! If I understand it correctly, the tests were performed using 55dB and 60dB attenuation buckets with weights [1.0, 0.5, 0.0]. This resulted in quite low recall of 47%, meaning that only half of relevant contacts with infected user were properly registered. Assuming current adoption rate of 20% this is further reduced to only 10% chance of being notified about the contact with infected person, assuming everyone is using the app properly all the time and always uploads keys.

After the results of this test RKI made a change of increasing bucket threshold from 60dB to 63dB, but this seems like a small modification. Were there any further tests to confirm how this change affected recall value? Perhaps third bucket could be used to register contacts also in the situations with heavy shadowing? This is related to "yellow card" proposal (https://github.com/corona-warn-app/cwa-app-android/issues/899#issuecomment-663475550)

@kbobrowski kbobrowski added the question Further information is requested label Aug 1, 2020
@SebastianWolf-SAP
Copy link
Member

Sorry for not replying earlier - we just reached out again to Fraunhofer IIS for additional information, as also mentioned in #312 (comment).

Mit freundlichen Grüßen/Best regards,
SW
Corona Warn-App Open Source Team

@ghost
Copy link

ghost commented Aug 25, 2020

Hello @kbobrowski,

thanks for starting the discussion. I know it has been a while since you have asked, and therefore I am happy to get back to you today with an answer from the RKI.
Regarding your question, about performing additional tests:
We have planed additional test series in September in cooperation with Frauenhofer IIS. In this series, we mainly will focus on Bluetooth signal reflection in metal tubes (bus, train and aeroplane simulations), as well as changes due to the Google and Apple ENF API versions (Version 1.5 and 1.6). Still, it will be an aspect that we examine the effects and impacts of the bucket threshold increment from 60dB to 63dB.

Thank you,
LMM

Corona-Warn-App Open Source Team

@kbobrowski
Copy link
Author

kbobrowski commented Sep 11, 2020

@SebastianWolf-SAP @GPclips thanks for keeping us in the loop. I just noticed (via community Slack) a blog post comparing Fraunhofer IIS study and study done by prof. Leith team:

In the last test series, using different mobile devices, about 80 percent of the encounters were recorded correctly. As such, we do not understand the results presented in the Irish study. As far as the Federal Ministry of Health is aware, the Irish study hasn’t been subject to a scientific independent peer review process.

Using accuracy value here (which was 79% in Fraunhofer IIS study) may be misleading, as this was a study with highly uneven class distribution (ground truth was 103 contacts with no exposure and 40 contacts with exposure).

Even if CWA did not work at all (no contacts registered), the accuracy value would be 103 / (103 + 40) = 72%. Recall or F1 score provide less-misleading information about true performance of the system, and these values were lower (47% and 56%).

Would be great to have some comparison between the next Fraunhofer IIS study and prof. Leith study, using comparable statistical values. I would also suggest to refrain in public communication from quoting accuracy values without mentioning uneven class distribution, as it may easily mislead the reader, or at least quote also reference accuracy for non-functioning app (72%).

@pdehaye
Copy link

pdehaye commented Sep 14, 2020

cc @doug-leith

@pdehaye
Copy link

pdehaye commented Sep 14, 2020

(@kbobrowski SwissCovid seems to have suffered from the same problems in the public reporting of accuracy/recall DP-3T/bt-measurements#4 )

@kbobrowski
Copy link
Author

@pdehaye I think the problem with reporting accuracy is that many understand "accuracy" in this context as "how well the app detects exposures", not as "how well the app stays silent if there is no exposure + how well the app detects exposures" which is the correct interpretation.

But the main difference between prof. Leith study and Fraunhofer IIS study is that the former was in enclosed metal space (tram / bus), and the latter in the large open space - not really comparable.

@ghost
Copy link

ghost commented Sep 30, 2020

Hello @kbobrowski and community,

I just reached out to the RKI asking for the September test results. We will inform you as soon as we can and keep you in the loop.

Thanks,
LMM

Corona-Warn-App Open Source Team

@daimpi
Copy link

daimpi commented Nov 13, 2020

@GPclips are there any news wrt the Fraunhofer followup study you can share with us? 🙂

@heinezen
Copy link
Member

This issue is probably no longer relevant after ENFv2 was integrated into the app. I don't know if it would make sense to also discuss more recent tests on the ENA here and it would be best to create a new issue for that.


Corona-Warn-App Open Source Team

@daimpi
Copy link

daimpi commented Jan 19, 2021

@heinezen I am still interested to learn the results from the September Fraunhofer tests as the underlying technology on the fundamental level (BLE attenuation as distance proxy) didn't change between ENFv1 and ENFv2 and those test should at least provide some insight into that.

Additionally if there are any new test they would also be interesting of course 🙂.

@heinezen heinezen reopened this Jan 21, 2021
@heinezen
Copy link
Member

@daimpi I'll try to get a hold of the results :)


Corona-Warn-App Open Source Team

@Ein-Tim
Copy link
Contributor

Ein-Tim commented Apr 18, 2022

@heinezen (I know that you left the Open-Source-Team a while ago) were you Abel to get the results? If they are not available to the public / it is not planned to make them public, I suggest to close this issue.

cc @dsarkar

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

6 participants