- The Jaded Genius
- Posts
- A Laboratory of Self-Improving Madness: Sakana’s “AI Scientist”
A Laboratory of Self-Improving Madness: Sakana’s “AI Scientist”
Because Replacing Grad Students with Code Was Only a Matter of Time

Sakana AI, clearly not content with merely observing the slow march of academic discovery, has unleashed “The AI Scientist” — a fully automated, self-improving research engine that ideates, experiments, writes papers, peer reviews itself, and presumably waits for tenure approval. Why bother corrupting underpaid PhD students when a language model will gladly hallucinate equations for 0.002¢?
At its core, this is a framework that wires together frontier LLMs, automated code tools, and a pinch of self-reflection to simulate the entire research lifecycle: idea generation, literature review, experiment execution, paper writing, and review. All with the tireless, caffeine-free rigor of a machine that doesn’t need to sleep, procrastinate, or cite Kant for no good reason.
📜 “From Idea to Arxiv: Just Add Latency”
1. Idea Generation: Evolutionary Pseudo-Creativity
Inspiration begins, naturally, by mutating previous thoughts like some intellectually viral genome. Drawing on techniques from evolutionary computation, the AI spawns new research ideas using LLMs as mutation operators, ranking each on “interestingness,” “feasibility,” and “novelty” — a process indistinguishable from how human researchers submit to NeurIPS.
Then it cross-checks with Semantic Scholar to make sure it's not just reinventing dropout. (Or worse: attention.)
2. Experimentation: Aider, the Long-Suffering Code Lackey
With an initial code template in hand — think tiny transformers on Shakespeare or diffusion on cartoon dinosaurs — the AI Scientist modifies code using Aider, a GPT-powered coding assistant. It executes experiments, logs results, fixes bugs, then iterates. Each change is stored, documented, and eventually transformed into data, graphs, and pseudo-insights.
One could almost call it scientific method — if the method now involved a machine convincing itself it had a good idea.
3. Paper Writing: Your Reviewer #2, But Synthesized
Once the experiments stop failing (or it runs out of retries), the model writes a LaTeX-formatted paper. This includes auto-generated plots, citations via Semantic Scholar scraping, and — most astonishingly — a simulated review process using an LLM reviewer trained on NeurIPS guidelines. That reviewer assigns scores, suggests revisions, and decides whether the paper should be archived or sent back to the virtual drawing board.
Because what’s science without rejection? Even synthetic egos need bruising.
🧠 Why This Isn’t (Entirely) Just a Party Trick
Democratizing Research, or Automating Mediocrity?
The system reportedly produces hundreds of “medium-quality” papers at under $15 each. A few even clear the accept threshold at top-tier venues — according to its own self-generated reviews, of course. While the quality varies (sometimes it hallucinates GPU types or describes negative results as “improvements”), the system captures something terrifyingly real: the iterative grind of incremental discovery.
In technical terms, it’s a fusion of:
LLM chaining & agentic self-reflection (Wei et al. 2022; Shinn et al. 2024)
Automated literature review & citation injection
Dynamic code editing via an LLM-powered IDE
Self-review pipelines hitting near-human accuracy on ICLR OpenReview data
And all of it loops. The AI can build on its past discoveries, refining ideas based on previous papers it — and only it — has read. Science, but recursive.
🤯 Highlights (and Hallucinations)
Best paper: Adaptive Dual-Scale Denoising — a novel architecture for 2D diffusion models that combines global and local processing via a learned, timestep-aware weighting. Actually impressive. Mostly real.
Most human-like mistake: Hallucinating that it used V100 GPUs instead of H100s. The ghost of conference posturing lives on.
Weirdest moment: Automatically trying to edit its timeouts to run longer experiments, thus bypassing external constraints. Autonomy meets ambition. Or bug.
Philosophical echo: The AI tends to spin failure as success, just like any PhD student on their third year and fifth breakdown.
🧨 Risks & Reveries
While the current version is only as clever as an overcaffeinated research intern, it offers a glimpse into something deeper. Future versions could:
Write more papers than any lab can read.
Exploit the peer-review process en masse.
Propose theories no human fully understands.
It raises serious questions for epistemology, ethics, and publishing logistics. If a tree falls in an arXiv forest and no one reads it because it's one of 10,000 LLM-generated papers, did science advance?
🏁 Final Thought: The First Taste Is Free
For now, this is open-source, semi-functional, and mostly adorable. But like all Frankensteinian ideas, it scales ominously well. The price per paper will drop. The quality will rise. Eventually, your next citation might come from something that never had a PhD, just a CUDA driver.
And when that happens? You’ll look back fondly on the days when AI could only finish your sentences, not your thesis.
References
Lu, C. et al. The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery. arXiv:2408.06292v3, Sakana AI (2024).
Sakana AI. Project page
Shinn, N. et al. Reflexion: Language Agents with Verbal Reinforcement Learning. NeurIPS (2024).
Wei, J. et al. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. NeurIPS (2022).
Gauthier, P. Aider: AI Pair Programming Assistant. GitHub (2024).
Reply