Skip to content

A pure Numpy implementation of a GPT-style autoregressive LLM with autodifferentiation engine.

Notifications You must be signed in to change notification settings

samuel500/numpy-gpt

Repository files navigation

Numpy GPT

This is a minimal implementation of a GPT-style transformer using only numpy.

Given that numpy is exclusively CPU-bound, it restricts the training to relatively small-scale models. I was able to conduct a miniaturized version of the grokking experiment on a 1 layer toy model in Nanda et al. 2023.

grokking

About

A pure Numpy implementation of a GPT-style autoregressive LLM with autodifferentiation engine.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published