Bad Minima of Predictive Coding Energy Functions
Simon Frieder · Luca Pinchetti · Thomas Lukasiewicz
2024 Poster
in
Affinity Event: Tiny Papers Poster Session 1
in
Affinity Event: Tiny Papers Poster Session 1
Abstract
We investigate Predictive Coding Networks (PCNs) by analyzing their performance under different choices of activation functions. We link existing theoretical work on the convergence of simple PCNs to a concrete, toy example of a network - simple enough to explicitly discuss the fixed points in its training stage. We show that using activation functions that are popular in mainstream machine learning, such as the ReLU, does not guarantee the minimization of the empirical risk during training. We show non-convergence on an illustrative toy example and significant accuracy loss in classification tasks on common datasets when using ReLU compared to other activation functions.
Video
Chat is not available.
Successful Page Load