-
Notifications
You must be signed in to change notification settings - Fork 449
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
f16 with WGPU #597
Comments
Linked to gfx-rs/wgpu#4384 |
Does anyone know if the limited buffer size in wgpu will be alleviated eventually? Even if f16 gets supported the buffer size limits will still be a barrier to running large models. |
You can manually override the limits when selecting the device : https://github.com/burn-rs/burn/blob/ed255c5561b85876cf02cbc4d48f35e1f0d29ac0/burn-wgpu/src/context/base.rs#L228 The limits are low for compatibility reasons I think, but I can increase |
Ho yes I didn't know that, I'll make a PR soon. |
My bad, I accidentally tested with tch. |
I get the following panic when trying to run stable diffusion: `thread panicked at 'Error in Queue::submit: Validation Error Caused by: |
It may happen when you run out of memory. You can try to lower the It might increase the computing time, but it's probably negligible for a big model. Once again, a value that I'm not sure how we should set it 😅. |
Setting |
Yes we could do that for now. There is an issue to optimize the memory strategy: #582. |
There is a tweet saying "float16 in webGPU finally works now" https://twitter.com/nisten/status/1698796718840598850 Worth looking into this and see if we need to update anything. |
Feature description
It would be great if burn-wgpu supported f16. Is there a timeline for this?
Feature motivation
Large models such as Stable Diffusion exceed wgpu's maximum buffer size when using f32. f16 support would enable some such models to be run with the wgpu backend.
The text was updated successfully, but these errors were encountered: