Close Menu
    Facebook LinkedIn YouTube WhatsApp X (Twitter) Pinterest
    Trending
    • IdeaSpark Revolver S titanium screwdriver on Kickstarter
    • From eggs to avocados – Germany’s Orbem raises €55.5 million for AI-powered MRI expansion
    • 7 Best All-Clad Deals From the Factory Seconds Sale (2026)
    • Americans worry sports betting hurts integrity even as participation keeps rising
    • Best Home Ellipticals in 2026: Smash Your Health Goals With These Full-Body Workout Machines
    • From Vietnam Boat Refugee to Reliability Engineering
    • Does Calendar-Based Time-Intelligence Change Custom Logic?
    • The UK government is backing AI that can run its own lab experiments
    Facebook LinkedIn WhatsApp
    Times FeaturedTimes Featured
    Tuesday, January 20
    • Home
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    • More
      • AI
      • Robotics
      • Industries
      • Global
    Times FeaturedTimes Featured
    Home»Tech Analysis»Nvidia Rubin’s Network Doubles Bandwidth
    Tech Analysis

    Nvidia Rubin’s Network Doubles Bandwidth

    Editor Times FeaturedBy Editor Times FeaturedJanuary 10, 2026No Comments5 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email WhatsApp Copy Link

    Earlier this week, Nvidia surprise-announced their new Vera Rubin structure (no relation to the not too long ago unveiled telescope) on the Consumer Electronics Show in Las Vegas. The brand new platform, set to achieve prospects later this 12 months, is marketed to supply a ten-fold discount in inference prices and a four-fold discount in what number of GPUs it might take to coach sure fashions, as in comparison with Nvidia’s Blackwell structure.

    The standard suspect for improved efficiency is the GPU. Certainly, the brand new Rubin GPU boasts 50 quadrillion floating-point operations per second (petaFLOPS) of 4-bit computation, as in comparison with 10 petaflops on Blackwell, no less than for transformer-based inference workloads like large language models.

    Nonetheless, specializing in simply the GPU misses the larger image. There are a complete of six new chips within the Vera-Rubin-based computer systems: the Vera CPU, the Rubin GPU, and 4 distinct networking chips. To attain efficiency benefits, the parts must work in live performance, says Gilad Shainer, senior vp of networking at Nvidia.

    “The identical unit linked differently will ship a very totally different stage of efficiency,” Shainer says. “That’s why we name it excessive co-design.”

    Expanded “in-network compute”

    AI workloads, each coaching and inference, run on massive numbers of GPUs concurrently. “Two years again, inferencing was primarily run on a single GPU, a single field, a single server,” Shainer says. “Proper now, inferencing is changing into distributed, and it’s not simply in a rack. It’s going to go throughout racks.”

    To accommodate these vastly distributed duties, as many GPUs as potential must successfully work as one. That is the intention of the so-called scale-up network: the connection of GPUs inside a single rack. Nvidia handles this reference to their NVLink networking chip. The brand new line contains the NVLink6 change, with double the bandwidth of the previous version (3,600 gigabytes per second for GPU-to-GPU connections, as compared to 1,800 GB/s for NVLink5 switch).

    In addition to the bandwidth doubling, the scale-up chips also include double the number of SerDes—serializer/deserializers (which allow data to be sent across fewer wires) and an expanded number of calculations that can be done within the network.

    “The scale-up network is not really the network itself,” Shainer says. “It’s computing infrastructure, and some of the computing operations are done on the network…on the switch.”

    The rationale for offloading some operations from the GPUs to the network is two-fold. First, it allows some tasks to only be done once, rather than having every GPU having to perform them. A common example of this is the all-reduce operation in AI training. During training, each GPU computes a mathematical operation called a gradient on its own batch of data. In order to train the model correctly , all the GPUs need to know the average gradient computed across all batches. Rather than each GPU sending its gradient to every other GPU, and every one of them computing the average, it saves computational time and power for that operation to only happen once, within the network.

    A second rationale is to hide the time it takes to shuttle information in-between GPUs by doing computations on them en-route. Shainer explains this by way of an analogy of a pizza parlor making an attempt to hurry up the time it takes to ship an order. “What are you able to do when you had extra ovens or extra employees? It doesn’t enable you to; you can also make extra pizzas, however the time for a single pizza goes to remain the identical. Alternatively, when you would take the oven and put it in a automotive, so I’m going to bake the pizza whereas touring to you, that is the place I save time. That is what we do.”

    In-network computing will not be new to this iteration of Nvidia’s structure. In truth, it has been in widespread use since round 2016. However, this iteration provides a broader swath of computations that may be finished inside the community to accommodate totally different workloads and totally different numerical codecs, Shainer says.

    Scaling out and throughout

    The remainder of the networking chips included within the Rubin structure comprise the so-called scale-out community. That is the half that connects totally different racks to one another inside the information heart.

    These chips are the ConnectX-9, a networking interface card; the BlueField-4 a so-called information processing unit, which is paired with two Vera CPUs and a ConnectX-9 card for offloading networking, storage, and safety duties; and eventually the Spectrum-6 Ethernet change, which makes use of co-packaged optics to ship information between racks. The Ethernet change additionally doubles the bandwidth of the earlier generations, whereas minimizing jitter—the variation in arrival occasions of data packets.

    “Scale-out infrastructure must guarantee that these GPUs can talk properly so as to run a distributed computing workload and meaning I would like a community that has no jitter in it,” he says. The presence of jitter implies that if totally different racks are doing totally different components of the calculation, the reply from every will arrive at totally different occasions. One rack will all the time be slower than the remaining, and the remainder of the racks, stuffed with expensive gear, sit idle whereas ready for that final packet. “Jitter means shedding cash,” Shainer says.

    None of Nvidia’s host of latest chips are particularly devoted to attach between data centers, termed ‘“scale-across.” However Shainer argues that is the subsequent frontier. “It doesn’t cease right here, as a result of we’re seeing the calls for to extend the variety of GPUs in an information heart,” he says. “100,000 GPUs will not be sufficient anymore for some workloads, and now we have to join a number of information facilities collectively.”

    From Your Website Articles

    Associated Articles Across the Net



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Editor Times Featured
    • Website

    Related Posts

    From Vietnam Boat Refugee to Reliability Engineering

    January 20, 2026

    NASA’s Historic Test Stands Meet Their End

    January 20, 2026

    Are ‘tech dense’ farms the future of farming?

    January 20, 2026

    UK to consult on social media ban for under 16s

    January 19, 2026

    Google appeals landmark antitrust verdict over search monopoly

    January 19, 2026

    How crypto criminals stole $713 million

    January 19, 2026
    Leave A Reply Cancel Reply

    Editors Picks

    IdeaSpark Revolver S titanium screwdriver on Kickstarter

    January 20, 2026

    From eggs to avocados – Germany’s Orbem raises €55.5 million for AI-powered MRI expansion

    January 20, 2026

    7 Best All-Clad Deals From the Factory Seconds Sale (2026)

    January 20, 2026

    Americans worry sports betting hurts integrity even as participation keeps rising

    January 20, 2026
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    About Us
    About Us

    Welcome to Times Featured, an AI-driven entrepreneurship growth engine that is transforming the future of work, bridging the digital divide and encouraging younger community inclusion in the 4th Industrial Revolution, and nurturing new market leaders.

    Empowering the growth of profiles, leaders, entrepreneurs businesses, and startups on international landscape.

    Asia-Middle East-Europe-North America-Australia-Africa

    Facebook LinkedIn WhatsApp
    Featured Picks

    Today’s NYT Mini Crossword Answers for Jan. 17

    January 17, 2026

    Our Favorite Earbuds for Android Users Are $60 Off

    November 3, 2025

    Sydney cube satellite maker Waratah Seed wins US space Logie

    August 14, 2025
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    Copyright © 2024 Timesfeatured.com IP Limited. All Rights.
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us

    Type above and press Enter to search. Press Esc to cancel.