Skip to content

Helena-Yuhan-Liu/BioRNN_RichLazy

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

BioRNN_RichLazy

This repository (tested in PyTorch 1.10.2, Torchvision 0.4.2, Neurogym 0.0.2, and Numpy 1.18.1) conducts recurrent neural network (RNN) training across different initial hidden weight ranks and logs the associated laziness measures that indicate the amount of network adjustments over the course of training. The laziness of the learning regime is quantified using hidden weight change norm, representation alignment, and tangent kernel alignment [1-3].

Usage

The primary script is main.py. Helper functions for data storage and retrieval are located in file_saver_dumper.py. To execute the main script, use the following command:

python3 main.py

Under the default settings provided, the code should complete in less than 30 minutes.

References

[1] L Chizat, E Oyallon, and F Bach. On lazy training in differentiable programming. Advances in Neural Information Processing Systems, 32, 2019.

[2] T Flesch, K Juechems, T Dumbalska, A Saxe, and C Summerfield. Orthogonal representations for robust context-dependent task performance in brains and neural networks. Neuron, 110(7):1258–1270, 2022.

[3] T George, G Lajoie, and A Baratin. Lazy vs hasty: linearization in deep networks impacts learning schedule based on example difficulty. arXiv:2209.09658, 2022

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages