Skip to content
OVEX TECH
Technology & AI

Ex-OpenAI Researcher Launches AI That Researches AI

Ex-OpenAI Researcher Launches AI That Researches AI

Karpathy’s “Auto-Researcher” Sparks Debate on AI’s Recursive Self-Improvement

Andre Karpathy, a prominent figure in the AI community known for his work at Tesla and OpenAI, has released a new open-source project that is sending ripples through the industry. Dubbed “Auto-Researcher,” the tool is designed to autonomously conduct machine learning research, a move that some are hailing as a significant step towards the long-theorized concept of an “intelligence explosion.” While Karpathy himself has a knack for dramatic pronouncements, the underlying technology and its potential implications are drawing serious attention from researchers and enthusiasts alike.

The Genesis of Auto-Researcher

Karpathy, who now leads an AI education company focused on making large language model (LLM) development accessible, has previously released open-source models and training codebases that allow individuals to build their own versions of models like GPT. His latest release, however, takes this a step further by introducing a system that can actively improve itself. The project is not a massive undertaking in terms of code size, but its ambition is substantial: to enable an AI to conduct research and enhance its own capabilities.

The concept echoes theories like the “intelligence explosion,” popularized by figures such as Leopold Ashenbrener, a former OpenAI safety researcher. This hypothesis suggests a future where AI becomes so advanced that it can outperform humans in AI research, leading to a rapid, self-accelerating cycle of intelligence advancement. While hypothetical, the idea is a recurring theme in discussions surrounding artificial general intelligence (AGI) and artificial superintelligence (ASI).

How Auto-Researcher Works

At its core, Auto-Researcher provides an AI agent with a simplified machine learning training setup. The agent’s task is to experiment autonomously, modifying code, training for a set period, evaluating improvements, and deciding whether to keep or discard its changes. This iterative process is akin to a “survival of the fittest” for code, where successful modifications persist and unsuccessful ones are abandoned.

Karpathy frames this as a fundamental shift in how AI research is conducted. He envisions a future, humorously depicted in the project’s README file, where human “meat computers” are replaced by autonomous AI research agents. The current iteration of Auto-Researcher allows users to program these agents not by directly editing Python files, but by writing instructions in markdown files. These markdown files act as the context and directives for the AI agents, effectively allowing users to “program the program.”

The system typically involves a `prepare.py` file for setup, a `train.py` file that the AI agent modifies to improve training processes, and a `programm.md` file which contains the human-provided instructions and goals for the agent. The agents then iterate on the `train.py` file, experimenting with aspects like architecture, hyperparameters, and optimizers, within a fixed time budget (e.g., five minutes) to see how much improvement they can achieve.

Real-World Experiments and Results

While the project is experimental, early results are already proving compelling. Karpathy reported that after running Auto-Researcher for about two days, tuning a model called NanoGPT, the system identified 20 changes that improved its validation loss – a measure of how well the model performs on unseen data. These changes were additive and transferable to larger models.

One notable outcome was an 11% reduction in the time it took to train a GPT-2 model, dropping from 2.02 hours to 1.8 hours. This demonstrates that the AI can autonomously discover ways to optimize the training process, effectively improving the next generation of AI models. Karpathy expressed mild surprise at how well his initial, relatively naive setup performed, especially on a project he considered already well-tuned by human experts.

The Auto-Researcher is built upon Karpathy’s NanoGPT project, which itself is a simplified, single-GPU implementation of LLM training. This contrasts with the large-scale, distributed training setups used by major AI labs. The focus is on accessibility, allowing individuals with a single GPU to train small language models, such as a character-level GPT on Shakespearean text, or a basic transformer model in a matter of minutes. This educational aspect is key to Karpathy’s broader mission.

Why This Matters: The Potential for Recursive Self-Improvement

The significance of Auto-Researcher lies in its demonstration of recursive self-improvement, a concept where an AI system can iteratively enhance its own capabilities. While current applications are on a smaller scale, the principle is profound. Companies like Google DeepMind (with AlphaEvolve) and Sakana AI (with Darwin Girdle Machine) have explored similar ideas, but Karpathy’s release makes this technology more accessible.

The potential impact is multifaceted:

  • Accelerated AI Development: If AI can research and improve AI more efficiently than humans, the pace of technological advancement could dramatically increase.
  • Democratization of Research: Open-source tools like Auto-Researcher could empower a wider community to contribute to AI progress, moving beyond the confines of large corporate labs.
  • New Research Paradigms: The shift from human-led hypothesis generation and testing to AI-driven autonomous research could unlock novel approaches and discoveries.

The Future: Swarms of Collaborating Agents?

Karpathy is already contemplating the next steps, including the possibility of creating swarms of collaborating AI agents. This would involve multiple agents working in parallel on research tasks, sharing findings, and promoting the most promising ideas to larger scales. He envisions a system where humans might contribute occasionally, but the bulk of the research and development is handled by AI.

The idea of distributed, collaborative AI research is reminiscent of how large-scale open-source software projects or even decentralized networks operate. Karpathy is exploring how platforms like GitHub could facilitate such a distributed research effort, allowing individuals worldwide to contribute their compute power and insights towards a common goal of AI self-improvement.

This approach offers a potentially different path to an intelligence explosion than one originating from a single, powerful AI lab. Instead, it could be a more distributed, emergent phenomenon driven by a global community of AI agents. The implications, as Karpathy suggests, are “wild,” and the space is undoubtedly one to watch closely as AI continues its rapid evolution.


Source: this EX-OPENAI RESEARCHER just released it… (YouTube)

Leave a Reply

Your email address will not be published. Required fields are marked *

Written by

John Digweed

1,631 articles

Life-long learner.