Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Any way to access the FP32 checkpoints? #259

Open
veritas9872 opened this issue Jan 12, 2025 · 2 comments
Open

[Question] Any way to access the FP32 checkpoints? #259

veritas9872 opened this issue Jan 12, 2025 · 2 comments

Comments

@veritas9872
Copy link

veritas9872 commented Jan 12, 2025

Hello!
I am currently doing research into quantizing Llama models.
However, I have noticed that the Llama checkpoints made public are always in BF16, including the embeddings.
Although this may be suitable for most use cases, I am currently trying some research that may be more precision-sensitive than most works.
Is there any way to access the master FP32 weights? It is my understanding that all weights have their master copies in FP32 but have some operations that are computed in BF16. I would be most grateful if the FP32 checkpoints could be made accessible.

@d-kleine
Copy link

Why do not just use Huggingface for that? You can specify the dtype as FP32 there.

@veritas9872
Copy link
Author

The HuggingFace checkpoints are in BF16. Even when I load the results in FP32, they are simply BF16 checkpoints cast to FP32.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants