FFLib is a neural network library based on PyTorch [2] that aims to implement several different types of layers and networks based on the Forward-Forward algorithm [1]. The library also provides a suite of tools for training, validating, testing, debugging and experimenting with Forward-Forward-based networks. We aim to make this library as close as possible to the original design and structure of the PyTorch library.
Dependencies needed to run the code and/or the examples:
To get started with our library check out the examples in the ./examples
folder.
You can run the examples either from the CLI or from inside VSCode interactively.
We recommend running them from the CLI for the first time, so you can fix any dependency issues
and making interactive execution only afterwards.
The Forward-Forward Algorithm was introduced in Geoffrey Hinton's paper "The Forward-Forward Algorithm: Some Preliminary Investigations" [1] with the following abstract:
The aim of this paper is to introduce a new learning procedure for neural networks and to
demonstrate that it works well enough on a few small problems to be worth further investigation.
The Forward-Forward algorithm replaces the forward and backward passes of backpropagation by
two forward passes, one with positive (i.e. real) data and the other with negative data which
could be generated by the network itself. Each layer has its own objective function which is
simply to have high goodness for positive data and low goodness for negative data.
The sum of the squared activities in a layer can be used as the goodness
but there are many other possibilities, including minus the sum of the squared activities.
If the positive and negative passes could be separated in time,
the negative passes could be done offline, which would make the learning much simpler
in the positive pass and allow video to be pipelined through the network without ever
storing activities or stopping to propagate derivatives.
There are 3 different types of Forward-Forward-based Neural Networks implemented in the FFLib:
The basic example of a Neural Network based on the Forward-Forward Algorithm. In the example file the following network specifications are used:
- 2 Dense layers with 2000 neurons each
- Predicting the MNIST dataset with TryAll probe
- Batch Size of 128
When a dense layer accepts an input, it is first detached from the graph and normalized before applying the weight multiplication. Essentially, it's just a linear layer that gets a detached normalized input. The trick is that each layer optimizes its "goodness" to be positive for "positive" data and negative for "negative" data. In the case of FFNet, the input and output has to be combined and given as an input to the network (to the first layer). Since there's no output of the network, we have to try all inputs when predicting. In the FFLib, there are special classes called probes. We use the TryAll probe to try all possible one-hot labels.
FF+C is a type of hybrid Neural Network that uses both, the Forward-Forward Algorithm and ordinary backpropagation to avoid the need ot the TryAll probe. The FF part is used as a feature extractor and it is trained the same way as the original FF network. The Classifier is another dense layer that as input gets the concatenated activations from both of the FF Dense layers and tries to predict the one-hot label.
The Forward-Forward Recurrent Neural Network is the most complex type of FF network implemented in the FFLib. It is made of a special type of layer called FFRecurrentLayer. The example file contains an example usage of the FFRNN with 2 FF Recurrent Layers, both with 2000 neurons. These networks are quite large due to the fact that each layer has not only weights from the previous layer, but also backward weights from the next layer. These networks have to be trained with multiple frames per batch, thus requiring even more time for both, training and inference.
We really appreciate contributions from the community! We especially welcome the reports of issues and bugs.
However, one may note that since this library is currently being heavily developed, the API may drastically change and all projects depending on this library have to deal with the changes downstream. We will however try to keep these at minimum.
The main maintainer of this library is Mitko Nikov.
We are using poetry to manage, build and publish the python package.
We recommend downloading poetry and running poetry install
to
install all of the dependencies instead of doing so manually.
To activate the virtual env created by poetry, run poetry env activate
to get the
command to activate the env. After activation, you can run anything from within.
There are three things that we are very strict about:
Run the following commands in the virtual env to ensure that everything is according to the guidelines:
mypy . --strict
black .
pytest .
Guidelines are now checked using GitHub Workflows. When developing the library locally, you can install act to run the GitHub workflows on your machine through Docker. We also recommend installing the VSCode extension GitHub Local Actions to run the workflows from inside VSCode, making the process painless.
Example scenarios are also tested in GitHub Actions by running them from the CLI.
Here are a few guidelines to following while contributing on the library:
- We aim to keep this library with as little run-time-necessary dependencies as possible.
- Unit tests for as many functions as possible. (we know that we can't cover everything)
- Strict Static Type-checking using
mypy
- Strict formatting style guidelines using
black
- No recursion (at our abstraction level)
- Nicely documented functions and classes