Google Researchers unveiled a new artificial intelligence (AI) Architecture Last Week that Can Enable Large Language Models (LLMS) to Remember the long-term context of events and topics. A paper was published by the mountain view-based tech giant on the topic, and the resultars claim that ai models train trained using using this Architecture displayed a more “Human-Like” Memory Retention. Notably, Google Ditched the Traditional Transformer and Recurrent Neural Network (RN) Architectures to Devlop a new method to teach ai models how to remumber contextual information.
Titans can scale ai models’ Context window more than 2 million tokens
The lead researchr of the project, Ali Behrouz, posted About the new Architecture on X (Formerly Known as Twitter). He Claimed That The New Architecture Provides a Meta in-Context Memory with Attention that Teaches Ai Models How to Remember the information at Test-Time Compute.
According to google’s paper, which has been Published In the pre-print online journey Arxiv, The Titans Architecture Can Scale The Context Windows of Ai Models to Larger Than Two Million Tokens. Memory has been a tricky problem to solve for ai developers.
Humans Remember Information and Events with Context. If Someone Asked A Person About What He Wore Last Weekend, they would be able Ay, when asked a follow-up Question about why they wore a brown jacket and deenim jeans last week, the person would be able to contextualize it with all these short-term and long-term information.
AI models, on the other hand, typically use retrieval-Augmented Generation (RAG) Systems, modified for transformer and RNN Architecture. It uses information as neural nodes. So, when an ai model has been asked a question, it accesses the particular node that contains the main information, as well as the nearby nodes that might containal or related to information. However, Once a Query is solved, the information is removed from the system to save processing power.
However, there are two downsides to this. First, an AI Model Cannot Remember Information in the long run. If one wanted to ask a follow-up question after a session was over, one would have to provide the full context again (Unlike How Humans Function). Second, AI Models Do A Poor Job of Retrieving Information Involving long-term context.
With Titans AI, Behrouz and Other Google Researchers Southt to Build an architecture which enables ai models to demloop a long-term memory that can be contrasting Putationally optimized.
To this end, the resigned an architecture that encodes history into the parameters of a neural network. Three variants were used – Memory as context (MAC), Memory as Gating (MAG), and Memory as a Layer (MAL). Each of these variants is suited for particular tasks.
Additionally, Titans uses a new surprise-based Learning Systen, which tells ai models to remumber unexpected or key information about a topic. These two changes allow titans Architecture to Showcase Improved Memory Function in llms.
In the Babilong Benchmark, Titans (MAC) Shows Outstanding Performance, where it effectively scales to larger than 2m context window, Outperming large models models like-4, 4, Lalama 3 . pic.twitter.com/zdngmtgiow
– Ali behrouz (@behrouz_ali) January 13, 2025
In a separet post, behrouz claimed a 3 + rag, And LLAma 3 70B.
6