Kush Bavaria, Ornn
In this theCUBE + NYSE Wired segment from “AI Factories – Data Centers of the Future,” Nebius co-founder and CBO Roman Chernin sits down with theCUBE’s John Furrier at the New York Stock Exchange to unpack how AI factories are reshaping enterprise infrastructure and the future of data centers. Chernin outlines Nebius’ two-track strategy: a multi-tenant cloud built for developer experience and managed services, and large-scale, mostly bare-metal deployments for hyperscalers and AI labs. He discusses the significance of Nebius’ Microsoft deal (described as “up to $20B” and set to become one of the largest single-site GB300 deployments) as both an engineering milestone and a way to feed scale and cash flow back into the core cloud business. The conversation explores why enterprises want “the baby of supercomputer in the cloud,” marrying cloud flexibility with supercomputing efficiency to minimize time-to-value without sacrificing performance. Chernin details Nebius’ specialization in AI-centric workloads (large distributed training and inference at scale), a platform roadmap that moves beyond infrastructure into inference, fine-tuning and reinforcement learning as services, and a commitment to helping customers build on open-source models for control, cost and data leverage. He traces customer waves from foundational model builders to vertical AI companies and tech-forward enterprises, noting early traction with firms like Shopify and momentum in regulated sectors such as healthcare following Nebius’ compliance milestones. With roots in Yandex’s large-scale engineering culture and meaningful exposure to ClickHouse, Chernin also weighs in on the economics of AI-scale infrastructure (power and capacity as gating factors), hybrid orchestration and sovereignty, and why latency priorities vary by use case – from reasoning models to voice agents – as AI factories become the new unit of value in modern enterprise compute.