-
Notifications
You must be signed in to change notification settings - Fork 276
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Launch requirements #18
Comments
bump, would like to know the VRAM requirements of each model |
For interest, tested using my 3090ti w/ 24gb dedicated: The standard model (inference), immediately got cuda memory issue... Ran model in fp16 and runs! Seems to hover around 15gb (standard, 16fp). Not scientific test though haha. |
Here is one more anecdotal data point: Getting started with the Galactica language model - Prog.World says:
FWIW, they also say:
|
Haven't tested w/o using 16fp. Or had time to do more tests, but standard using 16fp is giving some fantastic results. I'd imagine basic will be similarly good, though. edit: article looks like a good set up guide tho - mentions couple issues I also had to work out |
Note that the "base" model works in a free colab notebook, after selecting Runtime/Change runtime time and picking "GPU". |
Would you be so kind and share an example of a colab notebook? |
@vladislavivanistsev Here is an example of a colab notebook that you should be able to run for free with a GPU runtime: galactica on Colab I also note, from the paper:
|
My experience is similar: the (standard, 16fp) model runs for me with no issues across two GPUs (one RTX 2080 Ti + one GTX 1080 Ti), using about 15gb total (8gb+7gb). Not scientific test either :) but wanted to mention in case it helps someone. |
Where do you specify that the model should be fp16 and not fp32? |
Ah
|
Please add a column listing the inference memory requirements for the models, so people can easier judge how much GPU RAM they need for the versions. |
Who can share what equipment specifications are needed for each of the model sizes?
The text was updated successfully, but these errors were encountered: