Papers
Topics
Authors
Recent
Search
2000 character limit reached

Self-orthogonalizing attractor neural networks emerging from the free energy principle

Published 28 May 2025 in q-bio.NC, cs.AI, cs.LG, and cs.NE | (2505.22749v1)

Abstract: Attractor dynamics are a hallmark of many complex systems, including the brain. Understanding how such self-organizing dynamics emerge from first principles is crucial for advancing our understanding of neuronal computations and the design of artificial intelligence systems. Here we formalize how attractor networks emerge from the free energy principle applied to a universal partitioning of random dynamical systems. Our approach obviates the need for explicitly imposed learning and inference rules and identifies emergent, but efficient and biologically plausible inference and learning dynamics for such self-organizing systems. These result in a collective, multi-level Bayesian active inference process. Attractors on the free energy landscape encode prior beliefs; inference integrates sensory data into posterior beliefs; and learning fine-tunes couplings to minimize long-term surprise. Analytically and via simulations, we establish that the proposed networks favor approximately orthogonalized attractor representations, a consequence of simultaneously optimizing predictive accuracy and model complexity. These attractors efficiently span the input subspace, enhancing generalization and the mutual information between hidden causes and observable effects. Furthermore, while random data presentation leads to symmetric and sparse couplings, sequential data fosters asymmetric couplings and non-equilibrium steady-state dynamics, offering a natural extension to conventional Boltzmann Machines. Our findings offer a unifying theory of self-organizing attractor networks, providing novel insights for AI and neuroscience.

Summary

  • The paper introduces a formalism using the Free Energy Principle to derive self-organizing attractor networks that achieve orthogonal states via Bayesian inference.
  • It employs a continuous Bernoulli distribution to bridge network architecture with stochastic Hopfield-like dynamics, optimizing predictive accuracy and memory capacity.
  • In silico simulations validate key capabilities including robust sequence learning, optimal parameter tuning, and resistance to catastrophic forgetting.

Self-orthogonalizing Attractor Neural Networks Emerging from the Free Energy Principle

Introduction

The paper presents a formalism grounded in the Free Energy Principle (FEP) to describe the emergence of attractor networks. These networks inherently perform Bayesian active inference, optimizing predictive accuracy and model complexity. The study demonstrates that these networks self-organize without explicitly defined rules, using simulations to highlight emergent properties such as orthogonal attractor states which enhance generalization and reduce model complexity.

The Free Energy Principle and Network Dynamics

The Free Energy Principle, foundational to this study, provides a framework for understanding self-organization in complex systems. It posits that any viable system must minimize its variational free energy to maintain its structural integrity over time. This leads to systems that naturally partition into internal, external, and Markov blanket states, facilitating a dynamic interaction that is mathematically framed within this paper. Figure 1

Figure 1: Schematic of a particular partition illustrating internal (μ\mu) and external states (η\eta), separated by a Markov blanket comprising sensory (ss) and active states (aa).

The core of this mechanism is the transformation of FEP into computational architectures that learn and adapt without explicit instructions. This is realized through the introduction of "deep particular partitions" which enables arbitrarily complex hierarchies within networks, conceptualizing each component as a Bayesian entity interlinked with others.

Network Parametrization and Emergence of Attractor States

The authors introduce an internal parametrization based on the continuous Bernoulli distribution, creating a network state dynamic representative of complex probabilistic interactions without explicit intervention. Figure 2

Figure 2: Parametrization of subparticles showing interaction via micro-scale boundary states, with internal states influencing one another through deterministic scaling.

This approach mathematically bridges the network's architecture with stochastic Hopfield-like dynamics, enabling the emergence of orthogonal attractor states through adaptive learning mechanisms. This orthogonality is central, optimizing model efficiency by minimizing redundancy in internal representations and potentially enhancing memory capacity significantly.

Bayesian Inference and Learning Dynamics

The inference and learning processes operate on the principles of variational free energy minimization. Local node-level inference is aligned with Bayesian updating, where biases are adjusted by integrating input from adjacent nodes, reflecting a locally precise yet adaptive inference mechanism. Figure 3

Figure 3: Free energy minimizing attractor network where inference and learning processes coalesce to form orthogonal basis of inputs.

These same principles extend to macro-scale inference due to the structure's inherent property of nesting particular partitions, encapsulating both local and global network dynamics within a single unifying framework.

In Silico Simulations

The implementation of four diverse simulations underpins this formalism:

  1. Orthogonal Basis Formation: Simulation demonstrates network's ability to form robust orthogonal attractor states from correlated inputs, highlighting enhanced retrieval and generalization. Figure 4

    Figure 4: Performance landscapes revealing optimal inference parameters yielding highly orthogonal attractors.

  2. Learning Regimes: Extensive variance in training precision and evidence strength reveals an optimal parameter space fostering both high recognition and generalization capabilities.
  3. Sequential Dynamics: Networks trained in specific sequences produce asymmetric coupling matrices encoding temporal dynamics, thus showcasing the system's potential for sequence learning.

(Figure 5)

Figure 5: Sequence learned through network dynamics indicating spontaneous recall capabilities via asymmetric coupling networks.

  1. Catastrophic Forgetting Resistance: Spontaneous network activity reinforces attractor states, mitigating traditional issues with memory in dynamic environments. Figure 6

    Figure 6: Demonstration of the network's resistance to forgetting, maintaining retrieval performance after extensive free-running periods.

Conclusion

The study illustrates a comprehensive architecture for self-organizing attractor networks via the Free Energy Principle. These systems exhibit remarkable properties, such as orthogonal attractors and robust Bayesian inference, presenting implications for both natural and artificial intelligence. The framework's scalability and generalizability propose potent avenues for future research aimed at leveraging these principles in larger, more complex environments.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 7 tweets with 287 likes about this paper.