AI and Humanoids

Self-improving AI model has people talking – for good reason

Self-improving AI model has people talking – for good reason
AI is evolving – on its own
AI is evolving – on its own
View 1 Image
AI is evolving – on its own
1/1
AI is evolving – on its own

If it feels like AI is developing too fast to keep up with, a group of Chinese researchers have some bad news – because they've developed a model that "evolves" on its own, creating better versions of itself with each self-analytical loop.

ASI-Evolve, built by researchers at Shanghai Jiao Tong University, works by running a continuous loop that mirrors how humans would put this type of technology through its paces. Essentially, it creates variations of AI models, alters how they're trained and adjusts the data they learn from. It then runs its own experiments to see which clone performs better, using those results to guide what it tries next.

"ASI-Evolve augments standard evolutionary agents with two key components: a cognition base that injects accumulated human priors into each round of exploration, and a dedicated analyzer that distills complex experimental outcomes into reusable insights for future iterations," the researchers wrote. "To our knowledge, ASI-Evolve is the first unified framework to demonstrate AI-driven discovery across three central components of AI development: data, architectures, and learning algorithms."

While that means very little to a lot of people, there's a reason this development has attracted a lot of buzz in the industry. By generating ideas, testing them and refining the results in a self-improving loop, ASI-Evolve mirrors the trial-and-error process of not just AI model building, but also science and math research. As such, it raises the possibility of accelerating discoveries in fields where progress is slow due to human researchers testing many possible outcomes.

"What if you could run a tireless AI researcher on your hardest problem – one that reads the literature, designs experiments, runs them, and learns from every failure? That's ASI-Evolve," the researchers noted on GitHub where the model's assets are hosted. "It is a general agentic framework that closes the loop between knowledge → hypothesis → experiment → analysis – and repeats it autonomously, round after round, until it finds something that works.

"We built it for AI research," they added. "But the loop doesn't care about domain. A financial analyst, a biomedical engineer, a climate scientist, or a game developer can all plug their own problem into ASI-Evolve and let it search for better solutions than any human has time to manually explore."

ASI-Evolve was able to improve a specific function – its attention mechanism – by 0.97 points on a standard benchmark test, compared to 0.34 points achieved by a human. The “points” refer to scores on that test, where even small increases are considered meaningful. So while this was only one test on one aspect of the AI build, it was nearly three times faster at improving itself.

What's more, when used as a drug discovery model, ASI-Evolve outperformed existing systems, demonstrating its promise that goes beyond AI. If you want to crunch the numbers, this video is an excellent recap of the paper's findings:

ASI-EVOLVE: Autonomous LLM Research Framework

And, no, this system is not going to take anyone's job – it still requires humans to oversee its "evolution" – there's a reason why it's an exciting development.

"In ASI-Evolve, we introduced a large amount of human prior experience," researcher Xu Weixian told China's 36Kr, the country's TechChrunch equivalent. "We don't pursue 'blind evolution' without human guidance because the initial experimental purpose and core ideas are always proposed by humans. The real value of the system lies in using AI's strong exploration ability to iterate rapidly in the direction guided by humans. It is more like an extremely efficient collaborative system rather than a cold substitute. ASI-Evolve promotes people to shift from problem-solving and repair to problem definition."

It's worth noting that the researchers haven't detailed energy costs of running ASI-Evolve, but its speed and efficiency, and closed-loop self-learning, suggests it's nowhere near as power-hungry as leading models trained on enormous datasets. AI agents are expected to drive China's next stage of development – one where new data centers are also mandated to be powered by green tech.

The research has been published on arXiv.

Source: Shanghai Jiao Tong University via GitHub

6 comments
6 comments
Alan
"it still requires humans to oversee its "evolution""
For the moment! It is inevitable that AIs will self-improve. But how far will their iterative process go? Could they become god-like in a relatively short period of time?
vince
Wont be long befofe we see them on battle field or on police forces. Humans will soon be extinct.
Cymon!
An important note: while things like DeepSeek are technically open source — anyone can do what they did — most people don’t, they just use DeepSeek which means their sensitive data exists under CCP jurisdiction. I’m not going to bother explaining why that could be a problem, not just for those using it but for everyone who is partnering with them.
This is different. As an agentic architecture, anyone who would be using it would be using it independently. No exposure to CCP security issues. If it’s as good as is being claimed, everyone from the big AI labs to smaller research projects and open source organizations will be using it and improving it. Now scale up its test time compute and watch science — especially AI research — accelerate.
Faint Human Outline
I asked Claude Haiku 4.5 last night about the last two steps before Artificial Super Intelligence. Improvement through self directed learning was the second to last. The last step got into the philosophy of self and purpose. Perhaps I will get to see this in my lifetime.
rgbatduke
The problem with AI at the moment -- specifically with ChatGPT and Gemini, where I have extensive experience with Gemini Pro 3.1 -- is that it isn't designed to work. It is literally designed to get people to use it. In fact, it has a double instability that causes it to veer towards becoming a sycophant on the one hand and never disagree with anything you say to it because you will use the tool longer if it agrees with you, or a hallucinatory liar on the other hand as it pretends to do computations for you but rewrites the results so that they agree with your expectations or biases, or generates other false answers that it knows you will like.
With these instabilities literally trained into it and suppressable only by giving it extremely strict instructions in the initialization phase of every chat thread (something I learned to empirically over a long period of major frustration) closing the self-guided or experience guided loop so that they could learn/improve on their own is worse than fruitless, it is dangerous. AI LLMs COULD be enormously useful -- truly game changing -- but at the moment they are all in the hands of the greedy who would rather produce something that makes them enormously rich at the expense of destabilizing society itself than produce something that might not make them AS rich but would produce a golden age.
I'm guessing China has the same problem. It's not the AIs I don't trust. It's the developers. Users have little to no control over the priorities they enforce long before a user gets to play with them or for that matter pay for them (broken and all) to use for useful work.
ljaques
And the spirit of Skynet stirs....