You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello!
I am currently doing research into quantizing Llama models.
However, I have noticed that the Llama checkpoints made public are always in BF16, including the embeddings.
Although this may be suitable for most use cases, I am currently trying some research that may be more precision-sensitive than most works.
Is there any way to access the master FP32 weights? It is my understanding that all weights have their master copies in FP32 but have some operations that are computed in BF16. I would be most grateful if the FP32 checkpoints could be made accessible.
The text was updated successfully, but these errors were encountered:
Hello!
I am currently doing research into quantizing Llama models.
However, I have noticed that the Llama checkpoints made public are always in BF16, including the embeddings.
Although this may be suitable for most use cases, I am currently trying some research that may be more precision-sensitive than most works.
Is there any way to access the master FP32 weights? It is my understanding that all weights have their master copies in FP32 but have some operations that are computed in BF16. I would be most grateful if the FP32 checkpoints could be made accessible.
The text was updated successfully, but these errors were encountered: