• Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
2026 / 03 / 17 20:28 Tuesday
  • Login
mashdigi-Technology, new products, interesting news, trends
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
mashdigi-Technology, new products, interesting news, trends
No Result
View All Result
This is an advertisement.
Home Market dynamics

Google launches VaultGemma, the most powerful differentially private large-scale language model ever, marking a new milestone in balancing privacy and performance.

Author: Mash Yang
2025-09-14
in Market dynamics, Life, network, software
A A
0
Share to FacebookShare on TwitterShare to LINE

Google Research Team announces the launch ofNew VaultGemma Model, claiming to be the most powerful large-scale language model currently available, trained entirely from scratch and protected by differential privacy (DP). The model weights are simultaneously released to the Hugging Face and Kaggle platforms, allowing developers and academia to freely use, verify, and improve them.

Google launches VaultGemma, the most powerful differentially private large-scale language model ever, marking a new milestone in balancing privacy and performance.
▲Google Research Team announces the launch of a new VaultGemma model

As generative AI becomes more commonplace, privacy protection has become a crucial issue in AI development. Differential privacy reduces the risk of models memorizing individual data by introducing "noise" during training. However, this also presents challenges such as reduced training stability, increased batch size, and increased computational costs.

Google said that the research conducted in collaboration with DeepMind has established for the first time "differential privacy model scaling laws," which can accurately predict the optimal training configuration under different computing, privacy, and data budgets, becoming an important guide for training high-performance differential privacy models.

Google launches VaultGemma, the most powerful differentially private large-scale language model ever, marking a new milestone in balancing privacy and performance.
Differential PrivacyThe structure of the scaling law, Google points outPrediction loss can be accurately modeled using model size, number of iterations, and noise-to-batch ratio, simplifying the complex interactions between computation, privacy, and data budget.

VaultGemma is a new version based on Gemma 10, designed with a billion parameters. Through systematic experiments, the Google research team quantified the relationship between model size, number of training iterations, and noise ratio, and concluded that the optimal strategy for differentially private training is "smaller models with larger batch sizes." This strategy enables VaultGemma to achieve performance close to that of non-private models while maintaining higher privacy, with performance comparable to non-differentially private models from five years ago.

Technically, VaultGemma utilizes a scalable DP-SGD algorithm and an improved Poisson sampling method to ensure consistent batch size while maintaining strong privacy guarantees. The resulting model achieves sequence-level differential privacy (ε ≤ 2.0, δ ≤ 1.1e-10), ensuring that even if a single training example is queried, it is virtually impossible for the model to reproduce it. Google also conducted memorization tests, which showed that VaultGemma virtually "endorses" no training data.

Google launches VaultGemma, the most powerful differentially private large-scale language model ever, marking a new milestone in balancing privacy and performance.
▲The marginal benefit of increasing the privacy budget (epsilon) and computational budget (batch size) in terms of the impact of the noise-to-batch ratio

Google noted that while the performance of differentially private models is still slightly lower than that of fully non-private versions, the gap has been narrowed and there are clear research paths for further improvement. VaultGemma not only demonstrates Google's long-term commitment to privacy-preserving operations but also provides a reproducible and verifiable benchmark for industry and academia, driving the development of the next generation of privacy-focused AI.

Google launches VaultGemma, the most powerful differentially private large-scale language model ever, marking a new milestone in balancing privacy and performance.
▲Performance comparison of VaultGemma 1B (differential privacy) with its non-private version (Gemma3 1B), and an earlier baseline (GPT-2 1.5B).The results quantify the resource investment required for current privacy and show that modern differentially private training yields comparable utility to non-private models from about five years ago.

For developers, the release of VaultGemma not only provides pre-trained models but also comes with a comprehensive technical report and optimization recommendations, allowing businesses and research teams to tailor their models to their computing and privacy needs. This means that in the future, we can expect to see more businesses adopt AI with lower privacy risks, meet regulatory requirements, and protect user data, while still benefiting from high-performance models.

This is an advertisement.

Finally, Google emphasized that Vault Gemma is only the first step. In the future, it will continue to improve the differential privacy training mechanism, further enhance performance, and lower the computing threshold, so that "both safe and smart" AI will become the norm in the market.

Comparison of parameters and performance of VaultGemma, non-differentially private Gemma, and early GPT-2 models:

modelVaultGemma 1BGemma 3 1BGPT-2 1.5B
參數規模10 billion parameters10 billion parameters15 billion parameters
Privacy protectionDifferential privacy (ε ≤ 2.0, δ ≤ 1.1e-10)No differential privacyNo differential privacy
Training methodDP-SGD + Poisson Sampling OptimizationStandard non-DP trainingTraditional large batch non-DP training
Performance (relative to non-DP models)Close to the non-DP model 5 years ago (same level as GPT-2)Slightly higher than VaultGemmaLower than modern non-DP models, but close to VaultGemma
Data memory riskAlmost no memorization detectedThere is a certain risk of memorizationHigh risk of memoization (verified many times)
Release statusReleased, open source (Hugging Face & Kaggle)Released, open sourceHistorical models, publicly available for download
Tags: AIGemGoogleprivacyVaultGemmaArtificial wisdomprivacy
ShareTweetShare
Mash Yang

Mash Yang

Founder and editor of mashdigi.com, and student of technology journalism.

Leave a Reply Cancel Reply

The email address that must be filled in to post a message will not be made public. Required fields are marked as *

This site uses Akismet service to reduce spam.Learn more about how Akismet processes website visitor comments.

Translation (Tanslate)

Recent updates:

Almost imperceptible creases, stylus application a highlight: OPPO Find N6 Shenzhen on-site report, positioning itself before Apple's foldable iPhone debuts.

Almost imperceptible creases, stylus application a highlight: OPPO Find N6 Shenzhen on-site report, positioning itself before Apple's foldable iPhone debuts.

2026-03-17
Intel launches the Core Ultra 200HX Plus series of mobile processors! The Core Ultra 9 290HX Plus offers an even higher performance upgrade.

Intel launches the Core Ultra 200HX Plus series of mobile processors! The Core Ultra 9 290HX Plus offers an even higher performance upgrade.

2026-03-17
Micron launches a three-pronged attack! HBM4, SOCAMM2, and PCIe Gen6 SSDs enter full-scale mass production, becoming the "strongest support" for NVIDIA Vera Rubin.

Micron launches a three-pronged attack! HBM4, SOCAMM2, and PCIe Gen6 SSDs enter full-scale mass production, becoming the "strongest support" for NVIDIA Vera Rubin.

2026-03-17
mashdigi-Technology, new products, interesting news, trends

Copyright © 2017 mashdigi.com

  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Follow us

Welcome back!

Login to your account below

Forgotten Password?

Retrieve your password

Hãy nhập tên người dùng hoặc địa chỉ email để mở mật khẩu

Log In
No Result
View All Result
  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Copyright © 2017 mashdigi.com