AI and Humanoids

Powerful AI reasoning models can now run without giant data centers

Powerful AI reasoning models can now run without giant data centers
EPFL's Anyway Systems software offers a local alternative to remote power-hungry data
EPFL's Anyway Systems software offers a local alternative to remote power-hungry data centers
View 2 Images
EPFL's Anyway Systems software offers a local alternative to remote power-hungry data
1/2
EPFL's Anyway Systems software offers a local alternative to remote power-hungry data centers
This graphic illustrates the basic principle at work behind Anyway Systems
2/2
This graphic illustrates the basic principle at work behind Anyway Systems

Giant AI data centers are causing some serious and growing problems – electronic waste, massive use of water (especially in arid regions), reliance on destructive and human rights-abusing mining practices for the rare Earth elements in their components... And they also consume an epic amount of electricity – whatever else you might say about human intelligence, our necktop computers are super-efficient, using only about 20 Watts of energy.

They're also causing a massive supply-chain crunch for the high-powered GPUs required to run and train the latest models – there's no amount of "compute" that'll satisfy the endless appetites of these data centers and the companies pushing toward artificial general superintelligence.

One alternative to the Big Data model is running AI models locally – as in, right there on your own computer. But that limits the size and capability of the model you can run, and can put some harsh demands on your hardware. So what about distributing the task across small, local networks? At Switzerland’s École Polytechnique Fédérale de Lausanne (EPFL), a major technology and public research university, researchers have created software they’re now selling through their own company that takes out the middle-man of “Big Cloud.”

EPFL researchers Gauthier Voron, Geovani Rizk, and Rachid Guerraoui in the School of Computer and Communication Sciences, have announced Anyway Systems, an app that can be installed on networks of consumer-grade desktop PCs. There, Anyway downloads open-source versions of AI models such as ChatGPT, so you can ask questions globally, but process locally.

“For years,” says DCL head Rachid Guerraoui, “people have believed that it’s not possible to have large language models and AI tools without huge resources, and that data privacy, sovereignty and sustainability were just victims of this, but this is not the case. Smarter, frugal approaches are possible.”

Instead of using warehouse-bound arrays of servers, Anyway Systems distributes processing on a local network – in the above case with ChatGPT-120B, requiring a maximum of four computers – robustly self-stabilizing for optimal use of local hardware. Guerraoui says that while Anyway Systems is ideal for inference, it may be a bit slower responding to prompts, and “it’s just as accurate.”

ChatGPT-120B, for reference, is a powerful 'reasoning' model that comes in close behind OpenAI's o3 model on coding, math and health-specific benchmarks. It can use the web, write code in Python and perform complex chain-of-thought tasks, while running on relatively minimalist hardware.

Installing Anyway takes around 30 minutes. Because processing is local, users keep their private data private, and companies, unions, NGOs, and countries keep their data sovereign, and away from the clutches (and potential ethical compromises) of Big Data.

This graphic illustrates the basic principle at work behind Anyway Systems
This graphic illustrates the basic principle at work behind Anyway Systems

While home users would need several highly-specified PCs to form the local network needed to operate Anyway Systems, relatively small companies and organisations may well find they've already got everything they need. “We will be able to do everything locally in terms of AI,” says Guerraoui. “We could download our open-source AI of choice, contextualize it to our needs, and we, not Big Tech, could be the master of all the pieces.”

But doesn’t Google’s AI Edge already offer such abilities on a single phone?

“Google AI Edge is meant to be run on mobile phones for very specific and small Google-made models with each user running a model constrained by the phone’s capacity,” says Guerraoui. “There is no distributed computing to enable the deployment of the same large and powerful AI models that are shared by many users of the same organization in a scalable and fault-tolerant manner. The Anyway System can handle hundreds of billion[s of] parameters with just a few GPUs.”

According to Guerraoui, similar logic applies for people operating local LLMs such as msty.ai and Llama. “Most of these approaches help deploy a model on a single machine, which is a single source of failures,” he says, noting that the most powerful AI models require extremely expensive machines incorporating AI-specific GPUs like nVidia's 80 GB H100, which is the minimum price of entry for a single-machine installation of ChatGPT-120b.

And that's if you can get your hands on one – supply chain geopolitics, component complexity and rabid demand have combined to create lengthy waiting lists and crazy resale prices up to and beyond US$90,000 for the H100, a product that's supposed to retail for US$40,000.

Getting to that level of processing power is possible with consumer-grade PCs, but according to the EPFL researchers, you'd typically need “a team to manage and maintain the system. The Anyway System does this transparently, robustly and automatically.”

Ideas like Anyway might not put an end to the data warehouse model altogether, and it can't help with the energy-intensive task of training new models, but it does look like a neat and relatively low-lift way to bring high-powered, reasoning-capable AI models out of the Big Data ecosystem and into distributed local networks.

Sources: EPFL, Anyway Systems

Editor's note: this piece was substantially edited for additional context on January 5, 2026.

6 comments
6 comments
Pag
This article was clearly written by someone who knows next to nothing about LLMs. Removing the useless and mostly inaccurate fluff, it sounds like this team created a way to distribute models within a LAN instead of within a server rack. That could be interesting, but I'm worried that speed could take a big hit. Also, this won't run ChatGPT 5 or anything like it -- ChatGPT 120b is actually pretty small when it comes to AI models. However, being able to run a bigger local model on a couple of gaming PCs could be interesting to small teams and amateurs should speed hold up.
David Hanson
It took a few paragraphs to tell whether this was a news report or an editorial.
dave be
Was this article written by AI cause it doesn't seem to have a point or a point of view outside of promoting the business. It also makes it sound like they have something unique when distributed AI is a thing on pretty much all platforms and AI frameworks.
Marlen
I was thinking that this was maybe Folding at Home for AI? But I think it would only make sense with a distributed pricing model?
If this was built as something where you could: 1. purchase compute cycles form other users at a cost lower than current AI platforms, and 2. sell spare cycles on your machine to other users it could be a good way to reduce the cost of AI.
Since I only really use AI computation from 9-5 in my timezone, I could sell my extra compute for the other 16 hours of the day to users in other time zones. As long as the payout I was getting from the app covered the cost of electricity, cooling, and hardware wear, it could work?
HDBoomer
AI is next generation computer automation being (over)sold as nearly magical, and it is (unsurprisingly) no such thing. AI can misunderstand and misinform as well as any pathological liar; requires extraordinary amounts of, human, peer review.
Tech companies are trying to gain access to enormous amounts of water and electric power to build AI supporting, ginormous, data centers without addressing concerns about availability, consequences or cost.
These new data centers are being proposed in the desert, southwestern, US states; updating existing data centers/environments is finally getting any attention; GHG emissions will be increased by fossil fuel based, electricity generation; spent nuclear fuel rods will be produced; etc…
These are just some of the issues that those who want these new data centers are hiding behind NDA proliferation. After being built these data centers would have less than a few dozen, full time employees (I, remotely, supported network connectivity to multiple companies, data centers for 20+ years).
Tech swindlers are also overlooking the inability of LLM (Large Language Models), or any other AI variation, to tell truth from fiction, and therefore AI can create fictitious results in automation generated responses.
Automation, even when sold as AI, can be no better, or worse, than the humans who originated the software.
Thoughtful humans must apply the brakes to this AI rush, so that we need not find out, the hard way, that there are costs and consequences that must be considered, seriously considered, before proceeding with this push to put, “AI,” too close to everywhere, way too fast.
PAV
Isn't the idea of isolating your AI from the knowledge base that continuously updates counterintuitive to getting a complete answer to any question? Or is this more like trying to figure out a way to move a robot or something where it doesn't necessarily matter that they are in a closed environment?