# Navigate to this repo in the terminal
pip install -e .
from neural_nets.backprop import BackPropNet
-
Feedforward network of arbitrary size and activation functions
-
Backprop with arbitrary cost function
-
Feedback alignment based training for FNNs [1]
-
Recurrent network with arbitrary size and activtion functions
-
Backprop through time [WIP]
-
RNN traninng with feedback alignment [2]
Numpy and Matplotlib - planning to use JAX for some of the grad operations.
[1] Lillicrap, Timothy P., et al. Random synaptic feedback weights support error backpropagation for deep learning. Nature communications 7 (2016): 13276
[2] Murray, J. M. (2018). Local online learning in recurrent networks with random feedback. BioRxiv, 458570.