Building the Agentic Cloud
Can today’s cloud infrastructure support agentic AI, or will a new "agentic-native" approach to cloud soon be ushered in?
Written by Raphaelle d'Ornano | 6 min • July 07, 2025
Building the Agentic Cloud
Can today’s cloud infrastructure support agentic AI, or will a new "agentic-native" approach to cloud soon be ushered in?
Written by Raphaelle d'Ornano | 6 min • July 07, 2025
The artificial intelligence landscape stands at an inflection point that should profoundly reshape how we think about computational infrastructure. The emergence of agentic AI systems—autonomous agents capable of complex reasoning, coordination, and persistent memory—demands architectural approaches that differ from today's cloud computing paradigms.
This transformation raises critical questions about the $320 billion in projected infrastructure investment by hyperscale providers during 2025 and whether this massive spending targets the right architectural foundations. Infrastructure choices and transformation will determine which companies capture disproportionate returns, and how that may play out across three infrastructure categories: hyperscalers, neoclouds, and emerging agentic clouds.
Consider the scale of global infrastructure investment currently underway.
Amazon leads hyperscalers with $100 billion of investment announced for this year, followed by Microsoft at $80 billion, Google at $75 billion, and Meta at $60-65 billion. Spending this year by hyperscalers represents a 50% increase from 2024's already massive $197 billion deployment.
While these incumbents have a head start of years, if not decades, their supremacy seems at least partially vulnerable to brash neocloud upstarts such as CoreWeave.
In March, CoreWeave’s $23 billion valuation at IPO sparked celebration as a standout AI success. Dubbed the “AI hyperscaler” with NVIDIA backing, the company boasted $1.92 billion in revenue, an $11.9 billion OpenAI contract, and a $35 billion IPO target. Its higher GPU performance at lower cost appeared to cement its leadership in today’s AI infrastructure.
Yet as CoreWeave neared its IPO, there were growing doubts about the durability of its moat and model. The company adjusted its IPO targets in a challenging macro-economic context that did not help. After initial market volatility, CoreWeave’s stock has soared 316%, thanks in part to its recent announcement that the company had raised $7.5 billion in debt.
Amid this renewed euphoria, CoreWeave faces the strategic question of how to adapt to the agentic challenge. So far the focus has been on demand and AI infrastructure size.
Current cloud infrastructure operates according to principles that align well with traditional AI applications but create fundamental constraints for agentic systems.
The prevailing model is that of a traditional library system. Users request specific resources, receive them through centralized channels, and return them when finished. This model perfectly suits today's AI Agents, which are task-specific systems that operate with limited autonomy and reactive behaviors, aligning well with existing infrastructure designs prioritizing single-tenant GPU utilization and hub-and-spoke networking.
Agentic AI, however, demands what resembles a collaborative research university: multiple specialized entities working simultaneously, sharing persistent knowledge, and coordinating dynamically. This requires four critical architectural dimensions that existing providers handle poorly:
Multi-agent collaboration necessitates sophisticated orchestration where specialized agents—planners, retrievers, synthesizers—coordinate through persistent communication channels.
Dynamic task decomposition requires systems that adaptively redistribute computational workloads based on real-time changes.
Persistent memory architectures must support extended multi-agent workflows rather than stateless execution cycles.
Outcome-based resource allocation aligns infrastructure costs with successful task completion rather than raw computational consumption.
On top of this, agentic systems demand 3x to 4x the power of traditional AI agents to support persistent context storage and inter-agent coordination. This expansion affects power delivery, cooling infrastructure, and facility economics across multiple dimensions simultaneously.
The transition to agentic architectures demands evolution in both networking and storage paradigms at the same time—like transforming a centralized transportation system into an interconnected urban grid, where every node can communicate directly with many others.
Traditional load balancer designs used by hyperscalers cannot efficiently handle direct coordination protocols where agents maintain communication channels with multiple counterparts simultaneously. In addition, organizations must budget for explosive storage demands compared to traditional applications. This necessitates specialized approaches to data locality and caching strategies that optimize for task success rather than resource utilization—and a shift to coordination intelligence.
Protocol adoption accelerates these architectural demands with, for example, Anthropic’s Model Context Protocol (MCP). Adopted by Microsoft, OpenAI, and Google, it enables AI agents to interact with external systems. On the other hand, open-source MCP servers face security challenges as agent deployment scales across enterprise environments.
Unlike hyperscalers, neoclouds excel at powering current AI-driven workloads.
The long-term competitive position of neoclouds has depended on successfully adapting to emerging technological demands without abandoning the focused approaches that enabled their initial success. Neocloud success depends on transformation speed rather than capital scale alone, but their integration depth creates compound advantages across multiple dimensions simultaneously.
However, neoclouds are not necessarily optimal solutions for agentic paradigms.
"Unlike neoclouds that must abandon proven business models, hyperscalers can invest in agentic-native capabilities while maintaining existing revenue streams. "
Hyperscale cloud providers possess decisive structural advantages for developing agentic-native infrastructure through a combination of vertical integration, enterprise positioning, and capital deployment capabilities that smaller providers cannot replicate. Unlike neoclouds that must abandon proven business models, hyperscalers can invest in agentic-native capabilities while maintaining existing revenue streams.
Enterprise relationship leverage provides competitive protection through existing customer trust and procurement integration. Rather than requiring organizations to evaluate entirely new infrastructure platforms, hyperscalers extend established infrastructure partnerships into agentic capabilities through trusted channels. Microsoft's integration of MCP across multiple enterprise platforms exemplifies this approach, while Google leverages Workspace relationships and Amazon builds on AWS enterprise trust.
However, these advantages face important constraints that may limit adaptation speed. Hyperscaler scale creates organizational inertia that can slow architectural transformation, particularly when existing infrastructure investments represent substantial sunk costs. Their diverse customer bases with varying requirements may limit their ability to optimize specifically for agentic workloads without compromising performance for traditional applications.
While hyperscalers possess structural advantages for comprehensive agentic infrastructure, the emerging agentic cloud provider demonstrates the performance improvements possible with an architecture designed intentionally to address agentic requirements, rather than retrofitted to existing platforms.
Memory management optimizes specifically for persistent context storage across agent sessions. Networking infrastructure supports the low-latency agent communication patterns that coordination protocols require. Resource allocation systems enable outcome-based pricing models that align infrastructure costs with successful task completion.
Daytona Cloud represents one example of this emerging category, addressing critical requirements that traditional cloud platforms handle poorly. Similar specialized providers are targeting specific components of the agentic infrastructure stack, including observability tools for multi-agent monitoring and optimization platforms for agentic workload management.
The emergence of these purpose-built providers highlights the performance gaps in traditional infrastructure while demonstrating substantial market demand for specialized agentic capabilities that traditional infrastructure providers typically cannot deliver efficiently without fundamental architectural redesign.
The shift toward agentic systems introduces changes in economic models.
Traditional cloud economics operate on consumption-based pricing where infrastructure costs correlate with resource utilization—compute hours, storage capacity, network bandwidth—regardless of task completion success or coordination efficiency. Agentic systems, however, enable outcome-based pricing that aligns infrastructure costs with successful task completion rather than resource consumption alone. This transformation demands new business models from providers, with success metrics focused on task completion rates and agent coordination efficiency rather than traditional uptime or throughput measurements. It also creates opportunities for higher margins through architectures optimized for agentic workflows that deliver measurable business outcomes.
This economic transformation shapes infrastructure design priorities. Providers must optimize for task success over raw resource utilization, balancing performance, redundancy, and cost efficiency in ways that traditional cloud economics do not reward.
Pricing models will favor providers who can demonstrate measurable improvements in agent coordination capabilities, fundamentally redefining competitive benchmarks for infrastructure performance in agentic environments.
The agentic shift presents a fundamental challenge: building infrastructure for a technology whose full potential remains uncertain. ChatGPT emerged just 2.5 years ago, yet the industry is racing to lay tracks for a train whose destination remains unclear but continues gathering unstoppable momentum.
With 33% of enterprise software projected to incorporate agentic AI by 2028 and autonomous agents predicted to handle 15% of daily work decisions, infrastructure requirements will expand across all dimensions of agentic architecture, and the capital allocation challenge compounds this complexity.
Providers must simultaneously fund uncertain agentic technologies while preserving existing revenue streams, balancing innovation acceleration with operational stability during an extended transition period. Hyperscalers benefit from diversified portfolios that provide investment flexibility, yet all infrastructure providers will face pressure to commit substantial resources without clear short-term returns.
Organizations that approach this transition as an incremental upgrade rather than foundational transformation risk displacement by competitors who recognize that the future belongs to platforms designed for intelligence, not merely processing power. The agentic era demands infrastructure that thinks, coordinates, and adapts to dynamic requirements while maintaining the reliability and scale that enterprise applications require.
The providers that successfully navigate this architectural transformation will establish dominant positions in the next phase of cloud computing evolution, while those that optimize for yesterday's requirements risk obsolescence in tomorrow's agentic landscape.
This article originally appeared in Decoding Discontinuity, a Substack blog.