top of page

Decoupled Predictive Coding Networks for Backpropagation - PhD Proposal 2025

ABSTRACT: 


Backpropagation underpins modern deep learning but relies on point-estimate optimisation, providing no principled representation of epistemic uncertainty and limiting robustness and interpretability, while remaining vulnerable to distribution shift. Bayesian neural networks address some of these limitations but are difficult to scale, while predictive coding offers biologically grounded inference learning with explicit uncertainty at the cost of expensive iterative inference.
 

This PhD proposes a hybrid Bayesian learning framework that integrates predictive coding and backpropagation via a decoupled neural interface. A predictive coding network performs variational inference and generates oracle gradients corrected by a path–integral over inference dynamics, which are then amortised into a conventional backpropagation network. This yields an explicit variational free energy defined over backpropagation-trained models, enabling principled uncertainty
estimation, causal structure learning, and free-energy–based pruning while preserving computational efficiency.

 

The project will develop the theoretical foundations of this framework, implement and evaluate it on benchmark datasets, and explore extensions to neuromorphic hardware and active inference.

Downloads 

I'd like to offer for download my PhD proposal from 2025, as well as making the associated Github code available.


The general concept is establishing uncertainty over a backpropagation network, allowing for advanced concurrent structure learning. This is achieved using a decoupled neural interface (DNI) approach to synthetic gradient prediction. A PC network sharing the backpropagation network's parameters is used to generate these gradients. In order to allow for truncated inference ammortised synthetic gradients are fed through a path-integral. 

An extension, which is not yet extant in the code, but will follow, will involve the use of Kalman filters built on top of the path-integral to enact quadratic control over this trajectory an control variables. Subsequently (in continuous time) Kalman-Bucy filters may be used  to allow for full active inference across a backpropagated network.  

Thanks to Karl Friston, Christopher Buckley and Mehran Bazargani for their kind assistance with this project.

 

 

LINKS:

PDF research proposal
 

GitHub Code 

bottom of page