Applications of the Free Energy Principle to Machine Learning and Neuroscience
–arXiv.org Artificial Intelligence
In this thesis, we explore and apply methods inspired by the free energy principle to two important areas in machine learning and neuroscience. The free energy principle is a general mathematical theory of the necessary information-theoretic behaviours of systems which maintain a separation from their environment. A core postulate of the theory is that complex systems can be seen as performing variational Bayesian inference and minimizing an information-theoretic quantity called the variational free energy. The free energy principle originated in, and has been extremely influential in theoretical neuroscience, having spawned a number of neurophysiologically realistic process theories, and maintaining close links with Bayesian Brain viewpoints. The thesis is split into three main parts where we apply methods and insights from the free energy principle to understand questions first in perception, then action, and finally learning. Specifically, in the first section, we focus on the theory of predictive coding, a neurobiologically plausible process theory derived from the free energy principle under certain assumptions, which argues that the primary function of the brain is to minimize prediction errors. We focus on scaling up predictive coding architectures and simulate large-scale predictive coding networks for perception on machine learning benchmarks; we investigate predictive coding's relationship to other classical filtering algorithms, and we demonstrate that many biologically implausible aspects of current models of predictive coding can be relaxed without unduly harming the performance of predictive coding models which allows for a potentially more literal translation of predictive coding theory into cortical microcircuits. In the second part of the thesis, we focus on the application of methods deriving from the free energy principle to action. We study the extension of methods of'active inference', a neurobiologically grounded account of action through variational message passing, to utilize deep artificial neural networks, allowing these methods to'scale up' to be competitive with state of the art deep reinforcement learning methods.
arXiv.org Artificial Intelligence
Jun-30-2021
- Country:
- Africa > Mali (0.04)
- Asia > Middle East
- Jordan (0.04)
- Europe
- Finland > Uusimaa
- Helsinki (0.04)
- France (0.04)
- Germany > Hesse
- Darmstadt Region > Frankfurt (0.04)
- Switzerland > Zürich
- Zürich (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- Finland > Uusimaa
- North America > United States
- New York (0.04)
- Genre:
- Instructional Material (0.92)
- Overview (1.00)
- Research Report
- New Finding (1.00)
- Promising Solution (1.00)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Learning Graphical Models > Directed Networks
- Bayesian Learning (1.00)
- Neural Networks > Deep Learning (1.00)
- Reinforcement Learning (1.00)
- Statistical Learning (1.00)
- Learning Graphical Models > Directed Networks
- Representation & Reasoning
- Mathematical & Statistical Methods (1.00)
- Optimization (1.00)
- Uncertainty > Bayesian Inference (1.00)
- Machine Learning
- Information Technology > Artificial Intelligence