Co4 architecture

Co4 architecture: N denotes the number of input tokens, and each token has an embedding dimension of E. Q1, Q2,…,QL represent the latent query tokens input to the associated Q-TPNs. K1, K2,…,KN represent the Key tokens input to the associated K-TPNs. V1, V2,…,VN represent the Value tokens input to the associated V-TPNs. This configuration forms part of the “seeing” state (i.e., sensory processing).In the “seeing as” state (i.e., perceptual and interpretive state), triadic modulation loops among questions (Q), clues (keys, K), and hypotheses (values, V ) are executed through distal (D) and universal (U) contexts. Proximal (P) context represents normalization via information from neighboring neurons in the same population, including the prior information from the same neuron. The tpns associated with Q, K, and V are assumed to be analogous to three subtypes of pyramidal neurons, although their exact correspondence to neurobiologically distinguished subtypes is still under investigation. Through varying states of mind, high-level perceptual processing and wakeful thought, diverse, parallel reasoning chains are enabled. This mechanism incurs a computational cost of O(N · L),where L is a small fraction of the input length,making the overall cost approximately O(N).The triadic modulation loops, based on element-wise operations, add a nominal cost of L · N · E, which is significantly lower than that of the feedforward residual network used in standard Transformer blocks, a component Co4 does not require.Co4 can be viewed as a parallel, representation-level, silent yet deep form of Chain-of-thought (CoT) reasoning [56] (a quiet mind), enabling multi-perspective inference without requiring sequential token-level generation, much like the brain’s cortico-thalamic modulation. Credit: Ahsan Adeel.

The convergence of AI research and neuroscience is yielding exciting new pathways for innovation. By understanding the intricacies of the human brain, researchers can develop more refined and efficient AI systems. Conversely, AI models can provide valuable frameworks for understanding complex neurobiological processes.

Bridging the Gap: From Neurons to AI

Recent studies in neuroscience have focused on how mental state transitions influence interactions within layer 5 pyramidal two-point neurons (TPNs). Thes neurons exhibit distinct responses to external stimuli (receptive field) and internal states (contextual field). The processing of these inputs occurs at different sites within the neuron, offering clues for designing more nuanced AI models.

While current AI algorithms, including transformers, perceivers, and flamingo models, draw inspiration from the brain, they frequently enough fall short of replicating high-level perceptual processing and imaginative capabilities.Researchers are striving to create AI that more closely mirrors human cognition.

Co4: A Novel Approach to Cognitive Computation

Ahsan adeel,an Associate Professor at the University of Stirling,has been exploring ways to imbue AI models with higher mental states. His recent paper introduces Co4, a brain-inspired cognitive computation mechanism designed to mimic the dual-input, state-dependent processing observed in pyramidal tpns.

According to the paper, “Attending to what is relevant is fundamental to both the mammalian brain and modern machine learning models such as transformers. Yet determining relevance remains a core challenge, traditionally offloaded to learning algorithms like backpropagation. Inspired by recent cellular neurobiological evidence linking neocortical pyramidal cells to distinct mental states, this work shows how models (e.g., transformers) can emulate high-level perceptual processing and awake thought (imagination) states to pre-select relevant information before applying attention.”

The new model pre-selects relevant information before applying full attention, mirroring human perceptual reasoning. This involves a reasoning pattern based on questions, clues, and hypotheses, similar to human problem-solving.

“Triadic neuronal-level modulation loops among questions (Q), clues (keys, K), and hypotheses (values, V) enable diverse, deep, parallel reasoning chains at the representation level and allow a rapid shift from initial biases to refined understanding,” wrote Adeel.

The architecture achieves faster learning with reduced computational demand, at an approximate cost of O(N), where N is the number of input tokens. The model has shown promise in reinforcement learning, computer vision, and natural language question answering.

Implications for the Future of AI

The adapted transformer architecture was evaluated across various tasks, demonstrating its potential to enhance AI reasoning skills and bring them closer to human-like cognition.

Adeel concludes, “The initial evidence presented here is one of many reasons to believe that emulating the cellular foundations of higher mental states, ranging from high-level perceptual processing to deep, deliberate imaginative reasoning, could be a step toward cognitively meaningful machine intelligence.This approach opens the door not only to implementing large numbers of lightweight, inference-efficient AI modules, but also to moving these systems beyond mere information processing toward contextual reasoning, shifting from raw efficiency to real understanding.”

Understanding Transformer Architecture

Transformer architectures are a type of neural network that have revolutionized the field of AI,especially in natural language processing and computer vision.

At their core, transformers use a mechanism called “attention” to weigh the importance of different parts of the input data. This allows the model to focus on the most relevant information when making predictions. The Co4 model builds upon this foundation by incorporating brain-inspired mechanisms to improve the efficiency and accuracy of the attention process.