Backprop as Functor: A compositional perspective on supervised learning
Abstract
A supervised learning algorithm searches over a set of functions $A \to B$ parametrised by a space $P$ to find the best approximation to some ideal function $f\colon A \to B$. It does this by taking examples $(a,f(a)) \in A\times B$, and updating the parameter according to some rule. We define a category where these update rules may be composed, and show that gradient descent---with respect to a fixed step size and an error function satisfying a certain property---defines a monoidal functor from a category of parametrised functions to this category of update rules. This provides a structural perspective on backpropagation, as well as a broad generalisation of neural networks.
- Publication:
-
arXiv e-prints
- Pub Date:
- November 2017
- DOI:
- 10.48550/arXiv.1711.10455
- arXiv:
- arXiv:1711.10455
- Bibcode:
- 2017arXiv171110455F
- Keywords:
-
- Mathematics - Category Theory;
- Computer Science - Artificial Intelligence;
- Computer Science - Machine Learning
- E-Print:
- 13 pages + 4 page appendix