-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: use a very large coil dimension for 2D stacked NUFFT #39
Conversation
eadde29
to
0f96c5f
Compare
e9b9885
to
8380064
Compare
Alright some simple timing (H2D2H) reports the followings on a QUADRO P5000 with the data from the
Overall that's a 2x speed up on the Forward + Adjoint step (= a data consistency), it can be made faster by avoiding the roundtrip in the middle, and even more when cufinufft support the async copy / computations The stacked 2D Nufft on GPU also has a lower memory footprint (the oversampling grid is only 4 times bigger , not 8 times) making it possible to potentially have multiple operators running together (think fMRI) |
This PR improve the stacked nufft, by leveraging the coil dimension of the 2D operator.
For gpu backend which does asynchronous copy of coil data (e.g. only gpuNUFFT for now, cufinufft needs a few PR being merged upstream) this is great news, and potentially leads to increase in speed.
I still need to setup some rudimentary benchmark to compare between the two.