Large language models (LLMs) have already amazed us by reading, writing, and answering questions with impressive skill. But once their initial training is done, their knowledge tends to stay frozen, making it tricky to teach them new facts or skills — especially when we don’t have much task-specific data for retraining.
I recently came across MIT‘s new SEAL framework, an approach that flips that limitation on its head. Instead of relying on pre-designed training data and fixed instructions, SEAL lets AI models generate their own study notes and decide how best to train themselves. It’s a bit like how we humans prepare for tests — by rewriting notes, summarizing key ideas, and testing ourselves repeatedly, instead of just rereading textbooks.
How SEAL lets AI learn like a student
The core idea behind SEAL (which stands for Self-Adapting Large Language models) is that the AI produces short natural-language instructions called self-edits. These notes don’t just restate information but can infer new implications, summarize, or even suggest training tweaks like adjusting the learning rate. The AI then fine-tunes itself on these self-made notes, updating its internal parameters slightly.
Just like humans, complex AI systems can’t remain static for their entire lifetimes. They are constantly facing new inputs. SEAL aims to create models that keep improving themselves.
SEAL operates in two loops. In the inner loop, the model generates self-edits based on new readings and updates itself accordingly. Then it tests its own improvements by answering questions or solving puzzles. The outer loop uses reinforcement learning to keep only those self-edits that actually help performance — effectively teaching the AI how to write better notes over time.
Turning text into lasting knowledge
One of the coolest tests for SEAL was teaching the AI new factual knowledge. Instead of training directly on the original text, SEAL lets the model generate notes that highlight logical implications and key facts from a passage. Then the model trains on these notes using small updates.

Here’s where it gets interesting: without any adaptation, the model in the test answered about 33% of questions correctly. Training directly on the original passages barely bumped that up. But training on its own generated notes improved accuracy to nearly 40%. Even more impressive, notes generated by GPT-4.1 helped push accuracy to about 46%, while SEAL’s own self-learned notes nudged that further to 47%, surpassing the performance of a much larger model’s notes.
And this wasn’t just a fluke; SEAL kept its edge when learning from hundreds of passages simultaneously, suggesting it genuinely learned a general skill: how to write great study notes.
Adapting on the fly for problem solving
SEAL also shines on puzzle-like reasoning tasks that demand quick adaptation. Imagine a small AI given just a few examples to solve visual pattern puzzles with colored grids. Normally, without training, success was zero. With simple test-time training, it reached only 20%. After SEAL’s self-editing process rehearsed multiple study plans and picked the best, success jumped to over 70%!

This is a massive boost, showing how self-generated training strategies can help models adapt in real time to new challenges. While a human-designed ideal training plan still hits 100%, SEAL demonstrates that AI can develop its own clever study methods, cutting down the need for human-crafted solutions.

The challenges ahead and why this matters
Of course, SEAL isn’t perfect. One ongoing problem is catastrophic forgetting, where learning new information causes the model to gradually forget what it previously knew. The AI doesn’t crash outright, but older knowledge erodes as new self-edits overwrite it.
Also, running these self-edits requires fine-tuning and testing steps that take up to 45 seconds each, which could become expensive or slow with bigger models or massive datasets. Solutions like letting AIs generate their own tests to evaluate themselves might reduce this overhead in the future.

Despite the hurdles, SEAL points us toward a future where AI models don’t get stuck as static entities but instead keep growing, revising what they know and how they know it — much like how people learn throughout their lives. This capability would be a game changer for AI assistants that need to stay updated, scientific research bots that digest new papers, or educational tools that improve by catching their own mistakes and filling in gaps.
SEAL offers a concrete path toward language models that are not just trained once and frozen, but that continue to learn in a data-constrained world.
In other words, teaching AI to take and learn from its own notes might be the breakthrough needed for models that evolve continuously, making them more resilient, adaptable, and ultimately, smarter.
Key takeaways
- SEAL enables AI models to generate self-edits—study notes that help them improve continuously without human-designed datasets.
- Training on self-generated notes raised knowledge retention and reasoning success dramatically, showing models can learn how to learn.
- Challenges like catastrophic forgetting and costly training remain, but the approach points toward adaptable, lifelong learning AI systems.
It’s exciting to watch AI inch closer to learning more like we do – revising knowledge, testing itself, and growing over time instead of just stopping after initial training. SEAL is a step in that direction, and I can’t wait to see where this idea leads next.



