Attractor networks are a very general term in the domain of neural networks and dynamical systems. They refer to systems that evolve over time towards a set of stable states known as attractors. These attractors can represent memories, patterns, or other stable configurations in the system.

Many kinds of networks converge to attractors, the most common example being Hopfield Networks, which are recurrent neural networks that can store and retrieve binary patterns. In a Hopfield network, the attractors correspond to the stored patterns, and the network dynamics lead to convergence towards these patterns when presented with partial or noisy inputs.


Types

Point Attractors

Simplest case, tends towards a Fixed Point. A classic example of a single dimension fixed point attractor is the equation:

Theorem

Any dimensional damped linear system will have a fixed point attractor at the origin.

Line, Ring, and Plane Attractors

A natural extension of point attractors. Rather than a set of fixed points being the attractors, a whole continuous set of points form the attractor. For example, in a line attractor, the system can settle anywhere along a line in the state space.

A simple example would be that of any networking with neurons with which has an attractor Manifold defined by:

Is a line attractor network. In this case, attractor space itself is only one dimensional, even though the network state space is dimensional.

Simple dynamics definition of such a space(say 3 dimensions).

this yields a line attractor looking like the following:

First dimension means the system is stable on the line, second defines shape of the line, and third collapses higher dimensions proportionally to its own current value.

These kinds of attractors may not be straight lines. In fact, if the ends of line meet, it is now called a ring attractor.

Cyclic Attractors

Previous attractors, even line attractors, still converge on a single point and stay there. Cyclic attractors focus more on a convergence of a certain “flow” of a dynamical system. This is the limit cycle. A simple example would be a harmonic oscillator defined on an attractor Manifold:

is again defined over the attractor space, usually of a much lower dimension than the number of nodes in the network.

Forced Damped Pendulum is a chaotic dynamical system, however under certain conditions it can act as a cyclic attractor(in this case would be known as a limit cycle).

Chaotic Attractors

Also working over a “flow” or repeated traversal, but instead of converging on an exact flow, instead seems more like “bounded chaos”. It settle down to a obvious pattern, but the exact dynamics of said pattern are impossible to predict for long.

Classic example is the Lorenz Attractor, will still pull outward points into its cyclic bounds, but once there will follow a bounded flow that never converges to an exact cycle.


Biological Interpretations

What makes Attractor Networks important in Computational Neuroscience. We know that neural networks can implement any nonlinear dynamical system, so by extension this includes all attractor networks. One of the biggest goals of Computational Neuroscience is determining which attractors are important in the pursuit of explaining biological systems.