Nvidia GTC 2026: Wiwynn Showcases Nvidia Vera Rubin NVL72 AI Factory Infrastructure

Nvidia GTC 2026: Wiwynn Showcases Nvidia Vera Rubin NVL72 AI Factory Infrastructure

StorageNewsletter
StorageNewsletterMar 20, 2026

Key Takeaways

  • NVL72 integrates 72 GPUs, 36 CPUs, full liquid cooling.
  • Claims up to tenfold performance-per-watt improvement.
  • HGX Rubin NVL8 offers fanless 2U, 8‑16 systems per rack.
  • RTX PRO Server targets neural rendering with Arm CPUs.
  • Storage‑Next provides GPU‑directed 96‑drive NVMe, petabyte density.

Summary

At GTC 2026, Wiwynn unveiled a suite of Nvidia‑powered AI factory solutions built with Wistron, featuring the liquid‑cooled Vera Rubin NVL72 platform that unifies 72 GPUs and 36 CPUs. The announcement highlighted up to ten‑fold performance‑per‑watt gains, fanless HGX Rubin NVL8 modules, a compact RTX PRO Server for neural rendering, and the Storage‑Next GPU‑directed NVMe architecture. Wiwynn’s end‑to‑end integration promises faster time‑to‑value for data‑center operators seeking frontier AI training and inference capabilities. The offerings aim to accelerate AI adoption across global cloud infrastructures.

Pulse Analysis

The surge in generative AI and large‑language‑model workloads has forced data‑center operators to rethink traditional cooling and power architectures. Nvidia’s Vera Rubin platform, now paired with Wiwynn’s liquid‑cooled chassis, delivers a quantum leap in performance‑per‑watt, addressing the twin challenges of escalating energy bills and thermal constraints. By integrating 72 GPUs and 36 CPUs into a single rack‑scale unit, the NVL72 system reduces footprint while delivering the compute density required for next‑generation model training, positioning it as a cornerstone for hyperscale AI deployments.

Beyond raw compute, Wiwynn’s modular HGX Rubin NVL8 and RTX PRO Server expand the portfolio for diverse AI workloads. The fanless, 2U NVL8 module scales from eight to sixteen units per rack, leveraging Nvidia Spectrum‑X Ethernet or Quantum‑X800 InfiniBand for seamless bandwidth growth. Meanwhile, the ARM‑based RTX PRO Server targets specialized tasks such as neural rendering and AI‑driven design, offering a compact solution without sacrificing GPU performance. These offerings illustrate a strategic shift toward heterogeneous, liquid‑cooled infrastructures that can adapt to rapid workload fluctuations while maintaining high availability.

Storage‑Next completes the ecosystem by moving storage orchestration onto the GPU, enabling ultra‑high IOPS and sub‑millisecond latency across a 96‑drive NVMe array. This GPU‑initiated architecture not only accelerates inference for graph neural networks and retrieval‑augmented generation but also simplifies management through integrated telemetry and leak detection. Collectively, Wiwynn’s end‑to‑end integration with Nvidia’s cutting‑edge silicon equips enterprises with a future‑ready AI factory, reducing time‑to‑value and operational expenditure as the industry moves toward the agentic AI era.

Nvidia GTC 2026: Wiwynn Showcases Nvidia Vera Rubin NVL72 AI Factory Infrastructure

Comments

Want to join the conversation?