Nvidia had once again found itself at the bleeding edge of digital technology. At the glitzy keynote talk at Computex 2024 in Taiwan that announced the new platform, Nvidia’s CEO Jensen Huang tantalised the technology press with hints about the company’s next-generation AI-accelerating GPU platform, known to insiders as ‘Rubin’. More than just revealing a new product, the presentation announced a pattern: we could expect an upgraded version of ‘Rubin’ with enhanced AI capabilities to emerge in a year’s time. Then we could expect another upgrade a year later. If all goes to plan, Nvidia will have made an annual upgrade a part of the AI-technology life cycle. For a decade, analysts and industry insiders have greeted each successive revision of Nvidia’s GPU line with enthusiasm – and dismay.
Here lies the crux of Nvidia’s product strategy: underlying Huang’s tick-tock dance is a systematic, cadenced upgrade pattern designed to relentlessly advance the state of the art in accelerating AI. A year after he debuted the Blackwell Ultra, the company’s most powerful chip announced to date and due in 2025, Huang unveiled ‘Rubin’, set to debut in 2026. Under such a rhythm, Nvidia is assured not just of a frequent and regular drumbeat of innovation but its status as a world leader in the emerging market for AI accelerators.
The tech world held its breath as it got its first peek at what the Rubin AI platform could do: With onboard HBM4 high-bandwidth memory (3X the mobile bandwidth of the IBM POWER9 system) and the advanced NVLink 6 Switch (up to 3,600GBps) and acceleration as needed, Rubin represents Nvidia’s ability to make further breakthroughs in AI at a fraction of the cost in terms of both energy and dollars. After this, Nvidia announced ‘Rubin Ultra’.
Huang also announced ‘Vera’, an ARM CPU to supplement the Rubin GPUs on a new accelerator board nicknamed ‘Vera Rubin’, after the late American astronomer Vera Florence Cooper Rubin – drawing attention both to Nvidia’s branding and naming, and to their interest in luminaries of science and technology.
Nvidia’s public unveiling of Rubin, just three months after the announcement of Blackwell, is a calculated gamble. Nvidia generates its revenues via the design of its own chips rather than manufacturing them, and announcing the next generation of products while the current line is still operational creates the threat of the Osborne effect. Nvidia’s nerve in announcing Rubin prematurely is rooted in its confidence in its innovation pipeline and market dominance. It is an aggressive bet on being the company that will continue to change the game in AI technology, even when the board shifts underneath it.
When Nvidia announces the next chip or next AI accelerator platform, the world listens. And with an estimated 70-95 per cent market share in data centre GPUs, which power AI workloads, they have a bit of a lock on what the future of AI will look like – at least as far as the hardware goes. In a sector that changes fast, that has thousands of competing parts and companies, and which is racing to bring AI into every nook and cranny of the world, Nvidia can be forgiven if it is thinking big. But its reach in AI must match its grasp. And in a sector so hot, it certainly won’t have this market to itself for long. AMD and Intel and dozens of other companies are already racing to grab their share of the AI-accelerator market. A race for the future, if there ever was one, is about to get a lot more interesting.
Overall, Nvidia’s bold vision, instilled by the official Rubin AI platform announcement, highlights the company’s pivotal role in AI technological advancements. By effectively combining the right strategic planning, the resilience to innovate unapologetically, and a principle of broadening horizons in the science of AI, Nvidia remains at the helm leading the technorati into a new AI-driven future. From where we stand, one thing is certain: in an AI world, Nvidia isn’t just keeping up – they’re shaping AI’s future.
© 2024 UC Technology Inc . All Rights Reserved.