Replies: 6 comments 15 replies
-
I had created an issue about hardware requirements, maybe it will be possible to get answers there: #62 At the moment as it have to use ~300 GB model, looks like it is unlikely that it will work on CPU without changes to code and to the model, it also stated that it requires |
Beta Was this translation helpful? Give feedback.
-
I know this is obvious, but I have tried this on my 48 GB modified VRAM 3090, and it did not have enough vram. |
Beta Was this translation helpful? Give feedback.
-
Running a large language model (LLM) with the specifications you provided would require substantial hardware resources. Here's an overview of the hardware requirements you might need:
It's worth noting that running an LLM of this scale is a significant undertaking and would likely require a dedicated high-performance computing (HPC) cluster or a cloud-based solution from providers like AWS, Google Cloud, or Microsoft Azure. Additionally, you might need to consider techniques like model parallelism, tensor parallelism, or pipeline parallelism to distribute the workload across multiple devices effectively. |
Beta Was this translation helpful? Give feedback.
-
grok-1 GPU memory requirements I calculated those using the following parameters: Metrics Training | Memory in Gigabyte Metrics Inference | Memory in Gigabyte Finetuning Configuration | Metrics Finetuning | Memory in Gigabyte |
Beta Was this translation helpful? Give feedback.
-
So even RTX 4090 owners can not yet run GROK locally. Wonder how far can optimization go without considerable loss of quality... |
Beta Was this translation helpful? Give feedback.
-
What are the GPU RAM requrements please?
Can it be ported to CPU maybe?
Congratulations to all of us. Cautiously curious what will come out of this...
Beta Was this translation helpful? Give feedback.
All reactions