A minimal JAX/PyTorch approx-implementation of GPT based on karpathy's 'Let's build GPT'. The goal here is to learn frameworks (JAX, PyTorch), models (GPT, LLama, Gemma), evals (Hellaswag, MMLU) and more. The vision is to be able to train/finetune/infer SOTA small-medium models on (freely-available) TPUs.
-
Notifications
You must be signed in to change notification settings - Fork 0
A minimal PyTorch approx-implementation of GPT
License
alhaad/zeptogpt
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
A minimal PyTorch approx-implementation of GPT
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published