Most modern artificial intelligence models suffer from “catastrophic forgetting”: when trained on new data, they may distort or lose previously learned information. This limits their ability to be continuously and reliably updated for real-world applications. To address this, Google has introduced a new method called Nested Learning and its prototype architecture, Hope.
Google proposes viewing AI training not as a single continuous process, but as a hierarchical system of nested layers, each updated at different speeds:
• Upper Layer (Slow): Stores stable, long-term knowledge that rarely changes (e.g., grammar, core concepts).
• Lower Layer (Fast): Consists of flexible components that can be retrained quickly to absorb new or rapidly changing information.
In this framework, optimizers act like associative memory, helping the model recall previous corrections and avoid repeating the same mistakes.
Hope Prototype and Its Advantages
The prototype architecture Hope has already shown promising results in experiments. It:
• Demonstrated higher accuracy in language modeling,
• Showed better stability on tasks involving long context, compared to classical transformers.
If this technology proves successful, businesses could deploy more reliable AI agents that:
• Learn “on the fly” during interaction,
• Handle long context more efficiently without repeatedly reprocessing old data.
Google notes that Hope is still an early prototype and requires independent testing to confirm its viability.
ORIENT
