Skip to content

An implementation of nanogpt in jax from scratch based on Andrej Karpathy's tutorial series on Neural Networks.

License

Notifications You must be signed in to change notification settings

mahakal001/nanogpt-jax

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Nano GPT-jax

An implementation of nanogpt in jax from scratch ( Other than Optax for optimization and Equinox for handling PyTrees ) based on Andrej Karpathy's Let's build GPT Lecture.

Usage

  • The Shakespeare dataset is in data folder. You only need to configure hyper-parameters in nanogpt-jax/train.py as per your test settings and then run :
$ python train.py

TODOS

  • Write DropOut Layers.
  • LayerNorm.
  • Apply weight initializers.
  • Implement Adam.

References

About

An implementation of nanogpt in jax from scratch based on Andrej Karpathy's tutorial series on Neural Networks.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages