Skip to content

Latest commit

 

History

History
44 lines (25 loc) · 1.74 KB

README.md

File metadata and controls

44 lines (25 loc) · 1.74 KB

Can Functional Transfer Methods Capture Simple Inductive Biases? -- Code

This repository contains the implementation of all methods that we discuss in our publication Can Functional Transfer Methods Capture Simple Inductive Biases?

⚙️ Usage:

The experiments from that paper that use the methods from this repository can be found in orbit_transfer_recipes.

The experiments further require installation of:

⭐ Features:

Models:

models/cnn.py: Modular architecture of a plain CNN model

models/group_cnn.py: Implementation of a group-equivariant CNN model

models/group_equivariant_layers.py: Layers used in the G-CNN model

models/learned_equiv.py: Implementation of the Orbit model

models/mlp.py: Modular architecture of a plain MLP model

models/vit.py: Simplified implementation of a small VIT model

Transfer methods:

We implemented all transfer methods that we discussed in the paper as main-loop-modules (see nntransfer). This includes: attention transfer, knowledge distillation, representational distance learning and orbit transfer.

Simple MNIST-1D experiments:

A less modular and flexible implementation of all transfer methods and corresponding training can be found in trainer/simple_train.py and trainer/forward_methods.py. Fitting models are given in models/mnist_1d.py.

🐛 Report bugs

In case you find a bug, please create an issue or contact any of the contributors.