A person holding a smartphone displaying an AI chat interface in a high-tech office setting.

Shrinking the Giants: How Compact AI is Taking Back the Web

The world of artificial intelligence is currently hitting a massive wall. Right now, most of the AI we use depends on giant data centers that swallow up huge amounts of electricity and money. If the company running those servers has a bad day, your AI stops working. Even worse, some experts are worried that the companies providing all this computing power might start running out of space or breaking their promises.

Spanish startup Multiverse Computing thinks there is a better way. Instead of making AI bigger and hungrier, they are making it smaller and smarter. They are using a process called compression to take massive models from places like OpenAI and Meta and shrink them down. The goal is to make AI that runs directly on your own phone or laptop without needing an internet connection at all.

This shift is a huge deal for several reasons. First, there is the issue of privacy. When you use a typical chatbot, your data travels to a server owned by someone else. When the model lives on your device, your information never leaves your hands. Second, there is the cost. Running giant models is expensive. Smaller models are much cheaper to operate, which is why businesses are starting to pay very close attention.

Multiverse recently launched an app called CompactifAI to show off what they can do. Inside the app is a tiny model named Gilda. Gilda is small enough to live on your phone. If your device is powerful enough, Gilda answers your questions offline. If your phone is a bit older, the app uses a clever system to send the task to the cloud instead. They named this routing system after the One Ring from Lord of the Rings, which shows they have a sense of humor about how they control the flow of data.

But the real focus for Multiverse is not just a chat app. They are targeting big businesses that need AI in places where the internet is spotty or non-existent. Think about a drone flying over a forest or a satellite orbiting the Earth. You can’t exactly plug those into a giant data center in Virginia. They need AI that can think for itself right there on the hardware.

The company also just opened up a new portal for developers. This lets engineers grab these compressed models and build them into their own apps and tools. One of their latest models, HyperNova, is based on a massive OpenAI model but has been stripped down to run faster and cost less. In the world of coding, where speed is everything, these smaller models are starting to beat the giants at their own game.

Other companies are catching on too. Apple and Mistral are both working on their own versions of small, efficient AI. It is becoming clear that the future of tech isn’t just about who has the biggest computer. It’s about who can do the most with the least. Multiverse already has major clients like Bosch and the Bank of Canada, and they are reportedly looking to raise another 500 million Euros to keep pushing this tech forward. We are moving away from an internet where everything happens “out there” and toward a world where the smartest tools are the ones right in your pocket.