-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add deepspeed, xformers, kernl, transformerengine, ColossalAI, tritonserver, VoltaML, etc #21
Comments
Thanks for sharing this!
Our conclusion still holds for now and AITemplate is still the fastest. Please let us know if you have any other suggestions! We are looking for ways to improve this. |
Thanks for your detailed response :)
Is that true? They definitely talk about inference here (though I didn't explore too deeply to see what optimisations are applied): A bit further down on the page they reference some of the optimisations they make use of:
|
Another one I stumbled upon:
|
I've been bouncing around various StableDiffusion optimisations the last couple of weeks, and figured I would link out to some of the ones I remember in hopes that they can be explored/added into the benchmarks/comparisons here:
The text was updated successfully, but these errors were encountered: