Close Menu
    Trending
    • WNBA all-time great Tina Charles announces retirement
    • Justice Department seeks the names of 2020 election workers in Georgia’s Fulton County
    • Carney's fiscal update plays a charming tune, but falls off at the end
    • Ex-Ripple Exec Breaks Down The XRP To $10,000 Predictions, Is It Possible?
    • Aave says creditors are trying to seize stolen ETH before victims get their $71M back
    • Bitcoin Price Tops $81,000 For First Time Since January
    • Bose’s New Lifestyle Home Audio Lineup Includes A Speaker, Soundbar And Subwoofer
    • Three Myths that Derail Real Reform – The Health Care Blog
    FreshUsNews
    • Home
    • World News
    • Latest News
      • World Economy
      • Opinions
    • Politics
    • Crypto
      • Blockchain
      • Ethereum
    • US News
    • Sports
      • Sports Trends
      • eSports
      • Cricket
      • Formula 1
      • NBA
      • Football
    • More
      • Finance
      • Health
      • Mindful Wellness
      • Weight Loss
      • Tech
      • Tech Analysis
      • Tech Updates
    FreshUsNews
    Home » Nvidia Rubin’s Network Doubles Bandwidth
    Tech News

    Nvidia Rubin’s Network Doubles Bandwidth

    FreshUsNewsBy FreshUsNewsJanuary 10, 2026No Comments5 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Earlier this week, Nvidia surprise-announced their new Vera Rubin structure (no relation to the just lately unveiled telescope) on the Consumer Electronics Show in Las Vegas. The brand new platform, set to succeed in prospects later this 12 months, is marketed to supply a ten-fold discount in inference prices and a four-fold discount in what number of GPUs it could take to coach sure fashions, as in comparison with Nvidia’s Blackwell structure.

    The standard suspect for improved efficiency is the GPU. Certainly, the brand new Rubin GPU boasts 50 quadrillion floating-point operations per second (petaFLOPS) of 4-bit computation, as in comparison with 10 petaflops on Blackwell, at the least for transformer-based inference workloads like large language models.

    Nevertheless, specializing in simply the GPU misses the larger image. There are a complete of six new chips within the Vera-Rubin-based computer systems: the Vera CPU, the Rubin GPU, and 4 distinct networking chips. To realize efficiency benefits, the parts should work in live performance, says Gilad Shainer, senior vp of networking at Nvidia.

    “The identical unit related another way will ship a totally totally different degree of efficiency,” Shainer says. “That’s why we name it excessive co-design.”

    Expanded “in-network compute”

    AI workloads, each coaching and inference, run on giant numbers of GPUs concurrently. “Two years again, inferencing was primarily run on a single GPU, a single field, a single server,” Shainer says. “Proper now, inferencing is turning into distributed, and it’s not simply in a rack. It’s going to go throughout racks.”

    To accommodate these massively distributed duties, as many GPUs as potential must successfully work as one. That is the purpose of the so-called scale-up network: the connection of GPUs inside a single rack. Nvidia handles this reference to their NVLink networking chip. The brand new line contains the NVLink6 change, with double the bandwidth of the previous version (3,600 gigabytes per second for GPU-to-GPU connections, as compared to 1,800 GB/s for NVLink5 switch).

    In addition to the bandwidth doubling, the scale-up chips also include double the number of SerDes—serializer/deserializers (which allow data to be sent across fewer wires) and an expanded number of calculations that can be done within the network.

    “The scale-up network is not really the network itself,” Shainer says. “It’s computing infrastructure, and some of the computing operations are done on the network…on the switch.”

    The rationale for offloading some operations from the GPUs to the network is two-fold. First, it allows some tasks to only be done once, rather than having every GPU having to perform them. A common example of this is the all-reduce operation in AI training. During training, each GPU computes a mathematical operation called a gradient on its own batch of data. In order to train the model correctly , all the GPUs need to know the average gradient computed across all batches. Rather than each GPU sending its gradient to every other GPU, and every one of them computing the average, it saves computational time and power for that operation to only happen once, within the network.

    A second rationale is to hide the time it takes to shuttle knowledge in-between GPUs by doing computations on them en-route. Shainer explains this by way of an analogy of a pizza parlor attempting to hurry up the time it takes to ship an order. “What are you able to do in case you had extra ovens or extra staff? It doesn’t show you how to; you can also make extra pizzas, however the time for a single pizza goes to remain the identical. Alternatively, in case you would take the oven and put it in a automotive, so I’m going to bake the pizza whereas touring to you, that is the place I save time. That is what we do.”

    In-network computing will not be new to this iteration of Nvidia’s structure. Actually, it has been in frequent use since round 2016. However, this iteration provides a broader swath of computations that may be carried out throughout the community to accommodate totally different workloads and totally different numerical codecs, Shainer says.

    Scaling out and throughout

    The remainder of the networking chips included within the Rubin structure comprise the so-called scale-out community. That is the half that connects totally different racks to one another throughout the knowledge middle.

    These chips are the ConnectX-9, a networking interface card; the BlueField-4 a so-called knowledge processing unit, which is paired with two Vera CPUs and a ConnectX-9 card for offloading networking, storage, and safety duties; and at last the Spectrum-6 Ethernet change, which makes use of co-packaged optics to ship knowledge between racks. The Ethernet change additionally doubles the bandwidth of the earlier generations, whereas minimizing jitter—the variation in arrival instances of data packets.

    “Scale-out infrastructure must make it possible for these GPUs can talk effectively to be able to run a distributed computing workload and which means I want a community that has no jitter in it,” he says. The presence of jitter implies that if totally different racks are doing totally different components of the calculation, the reply from every will arrive at totally different instances. One rack will at all times be slower than the remainder, and the remainder of the racks, stuffed with pricey tools, sit idle whereas ready for that final packet. “Jitter means dropping cash,” Shainer says.

    None of Nvidia’s host of latest chips are particularly devoted to attach between data centers, termed ‘“scale-across.” However Shainer argues that is the following frontier. “It doesn’t cease right here, as a result of we’re seeing the calls for to extend the variety of GPUs in a knowledge middle,” he says. “100,000 GPUs will not be sufficient anymore for some workloads, and now we have to join a number of knowledge facilities collectively.”

    From Your Website Articles

    Associated Articles Across the Net



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWill Iranian Government Fall By February 2026?
    Next Article City pick up monster FA Cup victory
    FreshUsNews
    • Website

    Related Posts

    Tech News

    IEEE Smart Village Is Helping Electrify Rural Cameroon

    May 4, 2026
    Tech News

    DAIMON Robotics Wants to Give Robot Hands a Sense of Touch

    May 1, 2026
    Tech News

    AI Cyberattacks Meet Memory-Safe Code Defenses

    April 30, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    10 most intriguing players heading into 2025 college football season

    August 20, 2025

    Mindfulness, meditation, and the space in between

    October 25, 2025

    Ex Populus sues Elon Musk’s xAI over trade mark infringement

    August 25, 2025

    Arsenal hoping to tie up deal for Valencia defender Cristhian Mosquera

    June 26, 2025

    Minneapolis Mayor Frey: ‘Of course’ I’ll comply with DOJ probe

    January 18, 2026
    Categories
    • Bitcoin News
    • Blockchain
    • Cricket
    • eSports
    • Ethereum
    • Finance
    • Football
    • Formula 1
    • Healthy Habits
    • Latest News
    • Mindful Wellness
    • NBA
    • Opinions
    • Politics
    • Sports
    • Sports Trends
    • Tech Analysis
    • Tech News
    • Tech Updates
    • US News
    • Weight Loss
    • World Economy
    • World News
    Most Popular

    WNBA all-time great Tina Charles announces retirement

    May 5, 2026

    Justice Department seeks the names of 2020 election workers in Georgia’s Fulton County

    May 5, 2026

    Carney's fiscal update plays a charming tune, but falls off at the end

    May 5, 2026

    Ex-Ripple Exec Breaks Down The XRP To $10,000 Predictions, Is It Possible?

    May 5, 2026

    Aave says creditors are trying to seize stolen ETH before victims get their $71M back

    May 5, 2026

    Bitcoin Price Tops $81,000 For First Time Since January

    May 5, 2026

    Bose’s New Lifestyle Home Audio Lineup Includes A Speaker, Soundbar And Subwoofer

    May 5, 2026
    Our Picks

    Ethereum Sharp Exchange Outflows Sparks A Historic Supply Squeeze, Here’s What It Means

    October 4, 2025

    Christian McCaffrey draws inspiration from training partner Bijan Robinson as the 49ers beat the Atlanta Falcons

    October 20, 2025

    Gambling concerns could be inspiring NBA’s anti-tanking efforts

    February 18, 2026

    Thanasis Antetokounmpo To Sign One Year Deal With Bucks

    August 31, 2025

    Ranked-choice voting worked in New York City. Why not Seattle?

    July 29, 2025

    Famine determined in parts of Gaza, 500,000 experiencing ‘catastrophic’ hunger: Report

    August 22, 2025

    Poland to close last Russian consulate over ‘unprecedented act of sabotage’ | News

    November 19, 2025
    Categories
    • Bitcoin News
    • Blockchain
    • Cricket
    • eSports
    • Ethereum
    • Finance
    • Football
    • Formula 1
    • Healthy Habits
    • Latest News
    • Mindful Wellness
    • NBA
    • Opinions
    • Politics
    • Sports
    • Sports Trends
    • Tech Analysis
    • Tech News
    • Tech Updates
    • US News
    • Weight Loss
    • World Economy
    • World News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 Freshusnews.com All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.