Roughly, torch-lab is a lightweight, local, open-source cousin to Weights & Biases. I'm using this codebase to explore fundamental questions of neural networks: how they work and how we may interpret them. See the interpretability work from Anthropic and Redwood Research for examples.
Here's an example colab demonstrating the basics.
It is not intended for wide use, but I do believe this project could inform a more robust framework!
torch-lab features:
- An experimentation setup that abstracts away busywork
- The Experiment class allows grouping of related runs: varying parameters across a grid
- One can define Observables, helping standardize measurement during training.
- Configuration of runs is made clear, all parameters tracked, RNG initialization standardized for reproducibility.
- Much of the file i/o and naming is handled automatically
- A live visualization setup to see training in real-time
- Some common model implementations: mostly toy models for experimentation, but also...
- A GPU-capable Docker environment to run Jupyterlab
A companion tool is the "Transformer Visor", which lets you visualize the training process for a small transformer model. See a live demo here.