A minimal JAX/PyTorch approx-implementation of GPT based on karpathy's 'Let's build GPT'. The goal here is to learn frameworks (JAX, PyTorch), models (GPT, LLama, Gemma), evals (Hellaswag, MMLU) and more. The vision is to be able to train/finetune/infer SOTA small-medium models on (freely-available) TPUs.
-
Couldn't load subscription status.
- Fork 0
A minimal PyTorch approx-implementation of GPT
License
Couldn't load subscription status.
alhaad/zeptogpt
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
About
A minimal PyTorch approx-implementation of GPT
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published