-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Restore GalSim unit tests #534
Comments
For what it's worth, I am seeing the
|
Well the My feeling is we ought to allow an override right now for the GalSim benchmarks, setting |
Definitely better than commenting out test cases, but not as good as having test cases that actually work. Sorry if this is a question already answered elsewhere, but I've been out of the loop: if noise biases GalSim results so badly, why do we believe our results on actual catalogs? #157 is a good idea, but it seems like it might take a while. Is there any particular reason not to go back to the way active pixels used to be selected before Jeff re-did it to use the image? (See my Dec 9th comment on #482.) Or am I misunderstanding the problem? |
Yes, after rebasing I now see the correct galaxy probability estimates. That's good! I don't quite see how #507 could have fixed it, though, do you @gostevehoward ? |
For what it's worth, I don't have a lot of context on this and don't know
the right answer; I currently don't have an opinion on either the new way
or the old way.
I did ask why this doesn't affect SDSS/stripe82 in #482 and Jeff said (see
last comment):
I think it does affect SDSS and stripe 82, but not that much, in part
because an 8x8 grid around the center of the source always gets included no
matter how bright the light source is. Also, the noise_fraction is set
pretty low--iirc any pixel 50% higher than the background gets included.
Many sdss sources are quite a bit brighter than that.
…On Wed, Jan 18, 2017 at 3:42 PM Ryan ***@***.***> wrote:
Definitely better than commenting out test cases, but not as good as
having test cases that actually work. Sorry if this is a question already
answered elsewhere, but I've been out of the loop: if noise biases GalSim
results so badly, why do we believe our results on actual catalogs?
#157 <#157> is a good
idea, but it seems like it might take a while. Is there any particular
reason not to go back to the way active pixels used to be selected before
Jeff re-did it to use the image? (See my Dec 9th comment on #482
<#482>.) Or am I
misunderstanding the problem?
—
You are receiving this because you were assigned.
Reply to this email directly, view it on GitHub
<#534 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABC_kBn_2vS1TAwjU9d-X4mPI47PxB58ks5rTqN1gaJpZM4LnPgB>
.
|
@jeff-regier , any reason these arguments about stripe 82 don't apply equally to GalSim? If not, why do you think the GalSim results are biased? (Disclaimer: I haven't run these tests myself, so I'm just taking everyone's word that they are failing because of |
I don't know if or why #507 "fixed" it, no. It might be that Celeste finds
one of two minima, one for the galaxy fit and one for the star fit, and
that which minimum gets found is very sensitive to minor details right now
on this case. That is, the problem is masked by #507 but not really fixed.
(Or maybe another commit slipped in close to #507; I didn't look.)
…On Wed, Jan 18, 2017 at 3:51 PM Steve Howard ***@***.***> wrote:
For what it's worth, I don't have a lot of context on this and don't know
the right answer; I currently don't have an opinion on either the new way
or the old way.
I did ask why this doesn't affect SDSS/stripe82 in #482 and Jeff said (see
last comment):
I think it does affect SDSS and stripe 82, but not that much, in part
because an 8x8 grid around the center of the source always gets included no
matter how bright the light source is. Also, the noise_fraction is set
pretty low--iirc any pixel 50% higher than the background gets included.
Many sdss sources are quite a bit brighter than that.
On Wed, Jan 18, 2017 at 3:42 PM Ryan ***@***.***> wrote:
Definitely better than commenting out test cases, but not as good as
having test cases that actually work. Sorry if this is a question already
answered elsewhere, but I've been out of the loop: if noise biases GalSim
results so badly, why do we believe our results on actual catalogs?
#157 <#157> is a good
idea, but it seems like it might take a while. Is there any particular
reason not to go back to the way active pixels used to be selected before
Jeff re-did it to use the image? (See my Dec 9th comment on #482
<#482>.) Or am I
misunderstanding the problem?
—
You are receiving this because you were assigned.
Reply to this email directly, view it on GitHub
<#534 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABC_kBn_2vS1TAwjU9d-X4mPI47PxB58ks5rTqN1gaJpZM4LnPgB>
.
|
It's possible that To prevent any bias, when you call |
It's currently necessary to override this parameter to make galsim unit tests pass (see jeff-regier#534). This is the simplest way to code it but it's ugly in my opinion. We should discuss alternatives in the PR thread or an issue thread.
* permit overriding min_radius_px from top-level callers It's currently necessary to override this parameter to make galsim unit tests pass (see #534). This is the simplest way to code it but it's ugly in my opinion. We should discuss alternatives in the PR thread or an issue thread. * enable all galsim "unit test" cases now that active radius is expanded * updates calls to load_active_pixels! in unit tests
Fixed by #538 |
Right now, three out of the five cases exercised in
test_galsim_benchmarks.jl
are commented out because they're not working. I think I had all benchmark cases working well at some point but they weren't in unit tests so (as was inevitable) they got broken, and then when new cases broke they simply got commented out (broken window theory). We should get all five cases working again, and we shouldn't comment them out in the future, and maybe we should add more cases to the unit test.This will also be helped by adding better summary output to the GalSim benchmark script (probably as a result of #510). The point is to get people to notice when their changes degrade GalSim benchmark performance; currently that just doesn't happen.
@rgiordan
The text was updated successfully, but these errors were encountered: