
Image by Freepik
MIT’s SEAL Enables AI To Self-Teach And Continuously Adapt
MIT’s Improbable AI Lab researchers developed an innovative framework called SEAL (Self-Adapting LLMs). SEAL enables artificial intelligence systems to learn new skills after their initial training through self-teaching.
In a rush? Here are the quick facts:
- MIT developed SEAL, a framework that allows AI to teach itself new skills.
- SEAL rewrites its own training using self-generated summaries, quizzes, and notes.
- It achieved 40% better recall and 72.5% success in reasoning tasks.
The new system surpasses existing large language models, including ChatGPT, which the researchers explain generally remain fixed once trained and require extensive retraining to learn new information.
SEAL’s main breakthrough lies in its ability to let the AI generate its own training materials and then use these materials to fine-tune itself. Indeed, when presented with new data—such as a news article or example tasks—the system creates simplified explanations, related facts, or practice questions.
The researchers explain that this process mimics human learning methods, where people typically write notes and flashcards to improve their understanding and retention of new information.
During the “inner loop’’ phase, SEAL executes a small update known as a “self-edit.” The system conducts an “outer loop’’ assessment to verify the improvement of its performance after the update. When the AI identifies a positive change, it remembers the modification; otherwise, it attempts new approaches.
The researchers say that this iterative process allows the AI to continuously improve its knowledge and adapt to new challenges without the need for full retraining.
The researchers note that the method used by SEAL stands apart from all traditional reinforcement learning (RL) approaches. Agents in standard RL environments learn through trial-and-error to maximize their rewards when they interact with their surroundings.
SEAL uses RL as a tool to train its AI system to create and implement its own learning resources, which makes the language model as both an instructor and a student. The system produces “self-edits,” which include personalized instructions together with synthetic data for its refinement process. This approach enables persistent, self-directed updates that conventional RL methods, which do not adjust model parameters directly, cannot achieve.
MIT researchers tested SEAL on two fronts. First, in learning new facts, SEAL transformed raw text into implications and Q&A formats, resulting in a 47% accuracy improvement that surpassed even GPT-4.1’s training materials. Second, in abstract reasoning tasks, SEAL achieved a 72.5% success rate, outperforming models without reinforcement learning or standard training.
The researchers argue that the potential applications are wide-ranging. SEAL may enables various applications ranging from customized AI tutors, to self-evolving research collaborators, and autonomous agents that improve through experience.
The technology has the potential to transform enterprise AI systems by letting them develop internal thinking capabilities, which replace the need for costly retraining processes to handle fast data changes and user requirements.
However, challenges remain, for instance, SEAL can suffer from what the researchers call ‘‘catastrophic forgetting’’, where new learning overwrites previous knowledge, and the self-editing process requires significant computing resources.
To mitigate these issues, the team proposes hybrid systems that combine SEAL with external tools for temporary memory, reserving SEAL’s updates for essential, long-term knowledge.
Despite these hurdles, the MIT researchers believe that this technology will help machines achieve human-like adaptability, and lifelong learning abilities.