• Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
2026 / 01 / 25 09:12 Sunday
  • Login
mashdigi-Technology, new products, interesting news, trends
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
mashdigi-Technology, new products, interesting news, trends
No Result
View All Result
Home exhibition

NVIDIA's Rubin AI computing platform has entered mass production, with the powerful combination of Vera CPU and Rubin GPU reducing inference costs by 10 times.
DGX SuperPOD evolves again; Microsoft, AWS, and Google Cloud are among the first to adopt it.

Author: Mash Yang
2026-01-06
in exhibition, Market dynamics, Hard body, network, Processor, Topics
A A
0
Share to FacebookShare on TwitterShare to LINE

Following the Blackwell architecture, NVIDIA officially announced at CES 2026 this year.A new generation AI computing platform codenamed "Rubin"It has already entered mass production. NVIDIA CEO Jensen Huang emphasized that the Rubin platform was created to meet the needs of the next generation of AI factories, especially for complex tasks such as agentic AI, hybrid expert models (MoE), and long-context reasoning. Through the so-called "Extreme Codesign", the Rubin platform can reduce the token generation cost of AI inference by as much as 10 times.

NVIDIA's Rubin AI computing platform has entered mass production, with the powerful combination of Vera CPU and Rubin GPU reducing inference costs by 10 times.

Six core chips: Vera CPU and Rubin GPU lead the way

The core of the Rubin platform consists of six brand-new chips, among which the most eye-catching are the Rubin GPU and Vera CPU. The Rubin GPU is built using TSMC's 3nm process and has a built-in third-generation Transformer Engine. Its NVFP4 AI inference performance reaches 50 PFLOPS, which is 5 times that of the previous generation Blackwell architecture, and its training performance is also improved by 3.5 times.

The Vera CPU was designed to work with powerful GPUs. NVIDIA emphasizes that this is an Arm architecture CPU designed for AI inference, featuring 88 custom Olympus cores. Compared to the previous Grace CPU, the Vera CPU offers double the performance and boasts a memory transfer bandwidth of up to 1.2TB/s, enabling more efficient handling of large-scale data throughput.

High-speed interconnect: NVLink 6 and Spectrum-6

To enable these chips to work together, NVIDIA introduced the NVLink 6 Switch, providing up to 3.6TB/s of bandwidth per GPU, which is crucial for training large-scale MoE models. For network transmission, there are ConnectX-9 SuperNIC and Spectrum-6 Ethernet switches, supporting end-to-end connection speeds of up to 800Gb/s, ensuring high-speed data flow within the AI ​​factory.

DGX SuperPOD: The Infrastructure of an AI Factory

With the chip update, NVIDIA's supercomputer architecture DGX SuperPO has also ushered in a Rubin platform version.

• DGX Vera Rubin NVL72:This is a rack-level solution for extreme performance, integrating 8 systems in one rack, including 576 Rubin GPUs and 36 Vera CPUs, connected in series via NVLink 6, allowing the 576 Rubin GPUs to operate like a super-large GPU with a unified memory space, making it particularly suitable for processing very large models.

• DGX Rubin NVL8:For enterprises that require flexible deployment, NVL8 maintains a smaller liquid cooling specification and is paired with an x86 architecture CPU, allowing enterprises to more flexibly adopt the powerful computing power of Rubin computing.

NVIDIA's Rubin AI computing platform has entered mass production, with the powerful combination of Vera CPU and Rubin GPU reducing inference costs by 10 times.

Storage and Cybersecurity: Powered by BlueField-4 DPU

To address the key-value cache bottleneck during large model inference, NVIDIA introduced the Inference Context Memory Storage Platform based on the BlueField-4 DPU. This technology allows multiple GPUs to share context memory at high speed, improving inference speed and energy efficiency by 5 times.

In terms of cybersecurity, the Rubin platform also integrates cybersecurity solutions from partners such as Armis, Check Point, and F5, and uses BlueField DPU for real-time hardware acceleration protection to ensure the security of AI workloads.

NVIDIA's Rubin AI computing platform has entered mass production, with the powerful combination of Vera CPU and Rubin GPU reducing inference costs by 10 times.

Ecosystem support: All cloud giants join the effort

The NVIDIA Rubin platform has gained widespread support in the industry. Major cloud providers including Microsoft, AWS, Google Cloud, and Oracle have all announced their adoption of the Rubin system.

Microsoft will deploy the Vera Rubin NVL72 system in its next-generation "Fairwater" AI Gigafactory; CoreWeave, which focuses on AI computing power, will also be one of the first adopters.

Tags: BlackwellCESCES 2026DGX Rubin NVL8DGX Vera Rubin NVL72NvidiaRubyVera
ShareTweetShare
Mash Yang

Mash Yang

Founder and editor of mashdigi.com, and student of technology journalism.

Leave a Reply Cancel Reply

The email address that must be filled in to post a message will not be made public. Required fields are marked as *

This site uses Akismet service to reduce spam.Learn more about how Akismet processes website visitor comments.

Translation (Tanslate)

Recent updates:

Google and Apple are collaborating to release a feature that allows users to transfer photos stored in Google Photos to iCloud.

Can Google Photos also create memes? The new "Me Meme" feature uses generative AI to turn you into the protagonist of a meme.

2026-01-25
The smart glasses war has begun! Solos is suing Meta and Ray-Ban's parent company for patent infringement, seeking billions of dollars in damages and a sales ban.

The smart glasses war has begun! Solos is suing Meta and Ray-Ban's parent company for patent infringement, seeking billions of dollars in damages and a sales ban.

2026-01-25
TikTok's US operations have been split up! A new joint venture has been established, ByteDance's stake has decreased to 19.9%, and the algorithm will be licensed.

TikTok's US operations have been split up! A new joint venture has been established, ByteDance's stake has decreased to 19.9%, and the algorithm will be licensed.

2026-01-25
mashdigi-Technology, new products, interesting news, trends

Copyright © 2017 mashdigi.com

  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Follow us

Welcome back!

Login to your account below

Forgotten Password?

Retrieve your password

Hãy nhập tên người dùng hoặc địa chỉ email để mở mật khẩu

Log In
No Result
View All Result
  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Copyright © 2017 mashdigi.com