AI Podcasts
  • All Technology
  • AI
  • Autonomy
  • B2B Growth
  • Big Data
  • BioTech
  • ClimateTech
  • Consumer Tech
  • Crypto
  • Cybersecurity
  • DevOps
  • Digital Marketing
  • Ecommerce
  • EdTech
  • Enterprise
  • FinTech
  • GovTech
  • Hardware
  • HealthTech
  • HRTech
  • LegalTech
  • Nanotech
  • PropTech
  • Quantum
  • Robotics
  • SaaS
  • SpaceTech
AllNewsDealsSocialBlogsVideosPodcastsDigests

AI Pulse

EMAIL DIGESTS

Daily

Every morning

Weekly

Sunday recap

NewsDealsSocialBlogsVideosPodcasts
AIPodcastsInside the $41B AI Cloud Challenging Big Tech | CoreWeave SVP
Inside the $41B AI Cloud Challenging Big Tech | CoreWeave SVP
AI

Gradient Dissent

Inside the $41B AI Cloud Challenging Big Tech | CoreWeave SVP

Gradient Dissent
•January 6, 2026•53 min
0
Gradient Dissent•Jan 6, 2026

Key Takeaways

  • •CoreWeave optimizes GPU data throughput with custom storage and caching.
  • •Liquid‑cooled infrastructure boosts efficiency for high‑end AI GPUs.
  • •AI workload demand mirrors past analytics wave, needing best‑in‑class clouds.
  • •Microsoft and Google purchase services from CoreWeave as customers.
  • •Specialized design lets CoreWeave handle capacity bursts better.

Pulse Analysis

CoreWeave has built its business around the singular challenge of running AI training and inference at scale. By designing a purpose‑built object storage layer paired with its proprietary LottaCache, the company pushes as much data as possible directly into the GPU, eliminating the idle time that plagues generic public clouds. This focus on raw throughput, rather than a one‑size‑fits‑all API set, lets customers achieve higher performance per dollar on the most expensive component of any AI pipeline – the GPU itself. The result is a cloud offering that feels less like a shared service and more like a dedicated AI super‑computer.

The surge in generative AI mirrors the analytics boom of a decade ago, turning experimental models into business‑critical workloads that cannot tolerate latency or cost overruns. CoreWeave’s niche positioning satisfies enterprises willing to break from entrenched contracts in search of best‑in‑class performance. Even the giants of the industry—Microsoft and Google—have become customers, integrating CoreWeave’s specialized infrastructure into their own AI services. This partnership dynamic underscores a broader market shift: when the price of a GPU outweighs the convenience of a public cloud, organizations gravitate toward providers that can guarantee optimal GPU utilization and predictable pricing.

Technical differentiation comes from choices most hyperscalers avoid. CoreWeave’s data centers employ liquid‑cooled racks, a design that maximizes power density and reduces HVAC costs, enabling the deployment of the latest, power‑hungry GPUs. Coupled with a network architecture that processes inference calls inside the GPU, latency drops dramatically and the system can flexibly route traffic across multiple sites, improving availability during sudden demand spikes. As AI models grow larger and inference becomes ubiquitous, these engineering trade‑offs position CoreWeave to capture a growing slice of the $41 billion AI cloud market, while prompting larger clouds to reconsider their own specialization strategies.

Episode Description

The future of AI training is shaped by one constraint: keeping GPUs fed.

In this episode, Lukas Biewald talks with CoreWeave SVP Corey Sanders about why general-purpose clouds start to break down under large-scale AI workloads.

According to Corey, the industry is shifting toward a "Neo Cloud" model to handle the unique demands of modern models.

They dive into the hardware and software stack required to maximize GPU utilization and achieve high goodput.

Corey’s conclusion is clear: AI demands specialization.

Connect with us here:

Corey Sanders: https://www.linkedin.com/in/corey-sanders-842b72/

CoreWeave: https://www.linkedin.com/company/coreweave/

Lukas Biewald: https://www.linkedin.com/in/lbiewald/

Weights & Biases: https://www.linkedin.com/company/wandb/

(00:00) Trailer

(00:57) Introduction

(02:51) The Evolution of AI Workloads

(06:22) Core Weave's Technological Innovations

(13:58) Customer Engagement and Future Prospects

(28:49) Comparing Cloud Approaches

(33:50) Balancing Executive Roles and Hands-On Projects

(46:44) Product Development and Customer Feedback

Show Notes

0

Comments

Want to join the conversation?

Loading comments...