logoalt Hacker News

Nvidia Kicks Off the Next Generation of AI with Rubin

52 pointsby TSiegeyesterday at 5:45 PM39 commentsview on HN

Comments

mk_stjamesyesterday at 8:15 PM

Whenever I see press on these new 'rack scale' systems, the first thing I think is something along the lines of: "man I hope the BIOS and OS's and whatnot supporting these racks are relatively robust and documented/open sourced enough so that 40 years from now when you can buy an entire rack system for $500, some kid in a garage will be able to boot and run code on these".

show 2 replies
wmfyesterday at 8:32 PM

The blog post has more technical details and fewer quotes from customers: https://developer.nvidia.com/blog/inside-the-nvidia-rubin-pl...

show 1 reply
codybyesterday at 7:34 PM

If their new platform reduces inference token cost by 10x, does that play well or not well with the recently updated GPU deprecation schedules companies have been playing with to reduce projected cost outlays?

For context, my understanding is that companies have recently moved to mark their expected GPU deprecation cycles from 3 years to as high as 6 which has huge impacts on projected expenditures.

I wonder what the step was for the Blackwell platform from the previous. Is this slower which might indicate that the slower deprecation cycle is warranted, or faster?

show 3 replies
TSiegeyesterday at 5:45 PM

Extreme Codesign Across NVIDIA Vera CPU, Rubin GPU, NVLink 6 Switch, ConnectX-9 SuperNIC, BlueField-4 DPU and Spectrum-6 Ethernet Switch Slashes Training Time and Inference Token Generation Cost

Technical details available here https://developer.nvidia.com/blog/inside-the-nvidia-rubin-pl...

Groxxyesterday at 6:34 PM

... it took a couple searches to figure out that "extreme codesign" wasn't actually code-signing, but "co-design" like "stuff that was designed to work together"

show 3 replies
Animatsyesterday at 7:53 PM

Their own CPU, too - 88 ARM cores.

So it's an all-NVidia solution - CPU, interconnects, AI GPUs.

show 1 reply
exacubeyesterday at 8:25 PM

does anyone know how well this 5x petaflop improvement translates to real world performance?

I know that memory bandwidth tends to be a big limiting factor, but I'm trying to understand how this factors into it its overall perf, compared to blackwell.

metalliqazyesterday at 6:44 PM

Elon's emoji-filled blurb for that press release is the most cringe things I've seen this week.

show 3 replies
2OEH8eoCRo0yesterday at 8:03 PM

Rebuild all the data centers!

show 1 reply
dannersyyesterday at 7:23 PM

Riveting.