Menyala

Menyala is a Venture Studio headquartered in Singapore, with startups launched around the world. We bring together builders, ideas, network and capital to build disruptive ventures.

Featured

Revolutionizing AI Infrastructure: Unlocking the Power of Aggregated Compute at the Edge

Manoj Xavier
Menyala
Published in
5 min readFeb 21, 2025

--

The evolution of computing infrastructure has been a fascinating journey of decentralization and consolidation. In the 1990s, people ran their own servers, managing their computing needs independently. This era was marked by flexibility but also inefficiency, as individual setups were often underutilized and lacked scalability. The early 2000s saw the rise of centralized players like Amazon Web Services (AWS) and Google Cloud, who aggregated compute power into massive data centers, offering scalability and efficiency. However, this centralization brought new challenges — high costs, latency, and growing concerns over data privacy and vendor lock-in.

Today, we are witnessing the emergence of decentralized compute players such as Akash and io.net, heralding a new chapter in computing. These platforms leverage distributed computing resources across personal and enterprise devices to offer a flexible, cost-effective, and privacy-preserving alternative to traditional cloud infrastructure. This paradigm shift, powered by advances in edge computing and lightweight AI models, creates a unique opportunity to rethink AI infrastructure.

Enter aggregated compute at the edge: a model that taps into the latent compute capacity of everyday devices, enabling scalable, real-time, and secure AI applications

Market landscape of GPU and Storage compute networks

The Challenges of Centralized AI Infrastructure

Current AI infrastructure relies heavily on centralized cloud solutions, which come with several inherent challenges:

  1. High Costs: AI-focused GPUs are prohibitively expensive, often pricing out startups and small businesses.
  2. Privacy Concerns: Industries like healthcare and finance face significant risks when transferring sensitive data to cloud environments.
  3. Vendor Lock-In: Enterprises are tied to long-term agreements with major cloud providers, limiting flexibility and innovation.

These limitations not only hinder AI adoption for smaller players but also constrain innovation in industries that demand low-latency and secure AI solutions.

Reimagining AI with Aggregated Compute

The opportunity lies in creating a distributed compute network that taps into the unused processing power of personal and enterprise devices. This approach leverages everyday hardware — such as laptops, Raspberry Pis, and edge devices — to decentralize AI inference and deliver scalable, efficient solutions.

Key benefits of aggregated compute include:

  1. Cost Efficiency: By using existing latent compute resources, this model reduces reliance on expensive cloud GPUs, making AI accessible to startups and small businesses.
  2. Low Latency: Running AI models closer to the source of data (at the edge) minimizes delays, enabling real-time decision-making for applications like autonomous vehicles and predictive maintenance.
  3. Enhanced Privacy: On-device processing ensures that sensitive data never leaves its origin, addressing compliance requirements for regulated industries.

Applications Across Industries

Aggregated compute has the potential to revolutionize multiple sectors:

  • Healthcare: Deploying AI for medical imaging and diagnostics directly on secure, on-premise devices.
  • Retail: Real-time customer analytics and personalized recommendations in physical stores.
  • Finance: Fraud detection and compliance monitoring without exposing sensitive financial data to external systems.
  • Startups: Affordable AI tools for building innovative solutions without the overhead of cloud computing costs.

Interesting Use Cases

Building on the industry-wide applications of aggregated compute, edge optimization coupled with small language models (SLMs) introduces a transformative approach. SLMs, being lightweight and resource-efficient, excel in scenarios where bandwidth, compute power, or latency constraints make deploying traditional large language models (LLMs) impractical. This approach allows for faster inference, improved scalability, and tailored solutions for edge environments, making AI more accessible and cost-effective.

Use Cases:

  • Predictive Maintenance in Low-Resource Endpoints: Edge-optimized SLMs can process equipment sensor data locally to predict failures in real time, reducing downtime.
  • Hybrid Cloud Bursting for AI Scalability: Combine edge SLMs for real-time tasks with cloud LLMs for heavy computation during peak demand.
  • Enterprise Fleet Management: Enable on-device decision-making for logistics fleets using SLMs to optimize routes and monitor vehicle health, minimizing reliance on cloud connectivity.
  • Personalized Retail Experiences: Use SLMs in smart retail setups to provide localized, real-time product recommendations while respecting privacy and minimizing data transmission.

This integration not only enhances AI’s adaptability but also enables democratization by opening up new possibilities for hybrid, scalable, and efficient edge deployments.

Why Now?

Several factors make aggregated compute a timely innovation:

  1. Advancements in Edge AI: Smaller, more efficient AI models and improved edge hardware are making decentralized AI viable.
  2. Market Growth: Edge computing spending is forecasted to hit $232 billion by 2024, reflecting strong demand for distributed solutions.
  3. Open-Source Momentum: The rise of open-source AI models (e.g., Llama) provides developers with tools that can run efficiently on edge devices.

The Path Forward

Building an aggregated compute network requires a robust ecosystem of tools and partnerships to address challenges such as:

  • Scalability: Ensuring that distributed compute resources can handle growing AI workloads.
  • Reliability: Maintaining performance parity with centralized solutions.
  • Adoption Barriers: Overcoming vendor lock-in and convincing enterprises to adopt decentralized AI.

Despite these challenges, the potential of aggregated compute to democratize AI access and empower innovation makes it a compelling vision for the future.

Conclusion

Aggregated compute at the edge represents a transformative opportunity to address the cost, privacy and vendor lock-in challenges of traditional AI infrastructure. By unlocking the latent power of everyday devices, this decentralized approach can make AI more accessible, equitable, and impactful across industries.

As we stand at the crossroads of AI evolution, embracing distributed compute networks is not just a technological leap — it’s a necessity for building the next generation of scalable, efficient, and privacy-first AI solutions.

Are you an experienced entrepreneur or industry expert itching to build a startup to tackle this challenge? Menyala, a venture Studio headquartered in Singapore with startups launched around the world, brings together builders, ideas, network and capital to build disruptive ventures. Learn more at Co-found with us.

--

--

Menyala
Menyala

Published in Menyala

Menyala is a Venture Studio headquartered in Singapore, with startups launched around the world. We bring together builders, ideas, network and capital to build disruptive ventures.

Manoj Xavier
Manoj Xavier

Written by Manoj Xavier

Product Manager @ Menyala (Venture Building Studio founded by Temasek Holdings)

Responses (1)