Multiverse Computing Brings Compressed AI Models to the Mainstream

A Spanish startup is making smaller, more efficient AI accessible for on-device and enterprise use.

Multiverse Computing has launched an app and API portal for its compressed AI models, aiming to reduce reliance on external cloud infrastructure. These smaller models offer benefits like enhanced privacy and lower compute costs, especially for businesses. The move signals a shift towards more efficient and localized AI applications.

Katie Rowan

By Katie Rowan

March 20, 2026

4 min read

Multiverse Computing Brings Compressed AI Models to the Mainstream

Key Facts

  • Multiverse Computing launched an app (CompactifAI) and an API portal for its compressed AI models.
  • The compressed models are derived from major AI labs, including OpenAI, Meta, DeepSeek, and Mistral AI.
  • The CompactifAI app allows AI to run on user devices, enhancing privacy, but requires sufficient RAM and storage.
  • The primary target audience for Multiverse Computing's compressed AI models is businesses, not mass consumer adoption.
  • The API portal offers developers direct access to compressed models with real-time usage monitoring.

Why You Care

Ever worried about your personal data floating around in the cloud when using AI? What if your AI could run right on your phone, keeping your information private and secure? This is exactly what Multiverse Computing is working towards, pushing its compressed AI models into the mainstream. This creation could significantly change how you interact with artificial intelligence, offering more control and privacy.

What Actually Happened

Multiverse Computing, a Spanish startup, has recently made a significant move to democratize access to efficient AI. The company launched both an app, called CompactifAI, and an API portal, according to the announcement. These tools showcase and provide access to their compressed AI models. These models have been shrunk down from major AI labs like OpenAI, Meta, DeepSeek, and Mistral AI. The goal is to allow AI to run directly on a user’s device, reducing the need for large, external data centers.

The app offers a taste of ‘AI on the edge,’ meaning the processing happens locally. This keeps your data from leaving your device, enhancing privacy. However, there’s a catch: your mobile device needs sufficient RAM and storage. If not, the app intelligently switches to cloud-based models via an API. This routing is managed by a system named Ash Nazg, a nod to “The Lord of the Rings,” as mentioned in the release.

Why This Matters to You

This shift towards compressed AI models has practical implications for you and your business. Imagine running AI tasks without constant internet access or concerns about data breaches. This system offers a pathway to greater autonomy in how you use AI.

Here are some key benefits:

  • Enhanced Privacy: Your data stays on your device, avoiding cloud exposure.
  • Reduced Costs: Businesses can lower compute expenses by using smaller models.
  • Offline Capability: AI can function without an internet connection.
  • Increased Control: Developers gain direct access and transparency for production use.

For example, consider a small business that wants to use AI for customer support. Instead of paying for expensive cloud services, they could deploy a compressed AI model locally. This would handle inquiries efficiently and keep sensitive customer data within their own infrastructure. “The CompactifAI API portal [now] gives developers direct access to compressed models with the transparency and control needed to run them in production,” CEO Enrique Lizaso said in a statement. How might this newfound control change your approach to adopting AI solutions?

The Surprising Finding

What might surprise many is that while the CompactifAI app offers significant user benefits, its mass customer adoption isn’t the primary focus. The initial app usage figures, with only 2,000 global downloads on iOS and Android combined since its launch, suggest a limited direct consumer impact so far. This finding challenges the assumption that the goal is widespread individual user adoption. Instead, the real target audience for Multiverse Computing is businesses, as the company reports. The emphasis is on providing an API portal for developers. This allows enterprises to integrate these efficient models into their own operations. It seems the strategy is to empower other businesses to build with compressed AI, rather than directly serving millions of end-users with a standalone app.

What Happens Next

Looking ahead, Multiverse Computing’s focus will likely remain on expanding its enterprise offerings. We can expect more businesses to explore these smaller AI models in the coming months, particularly in late 2026 and early 2027. The company will probably refine its API portal based on developer feedback. For example, imagine a manufacturing plant using these models for predictive maintenance directly on factory floor equipment. This could prevent costly downtime without sending sensitive operational data to the cloud.

For readers, consider evaluating your current AI infrastructure. Are you overspending on cloud compute, or are privacy concerns holding you back? Investigating solutions like Multiverse Computing’s compressed AI could offer a more efficient path forward. The industry implication is clear: the demand for AI efficiency will continue to grow, pushing more companies towards localized and cost-effective solutions, as mentioned in the release.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice