• Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
2025/12/12 20:43 Friday
  • Login
mashdigi-Technology, new products, interesting news, trends
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
  • Topics
  • Artificial wisdom
  • Autopilot
  • network
  • Processor
  • 手機
  • exhibition activities
    • CES
      • CES 2014
      • CES 2015
      • CES 2016
      • CES 2017
      • CES 2018
      • CES 2019
      • CES 2020
    • MWC
      • MWC 2014
      • MWC 2015
      • MWC 2016
      • MWC 2017
      • MWC 2018
      • MWC 2019
    • Computex
      • Computex 2014
      • Computex 2015
      • Computex 2016
      • Computex 2017
      • Computex 2018
      • Computex 2019
    • E3
      • E3 2014
      • E3 2015
      • E3 2016
      • E3 2017
    • IFA
      • IFA 2014
      • IFA 2015
      • IFA 2016
      • IFA 2017
    • TGS
      • TGS 2016
  • About us
    • About mashdigi
    • mashdigi website contact details
No Result
View All Result
mashdigi-Technology, new products, interesting news, trends
No Result
View All Result
Home Life

NVIDIA reveals how scaling laws drive smarter, more powerful AI
Scaling laws describe how the performance of AI systems improves as more training data, model parameters, or computing resources are added.

Author: mashdigi news content
2025-02-17
in Life, Market dynamics, network
A A
0
Share to FacebookShare on TwitterShare to LINE

The development of AI to date has required three major laws to describe how different utilization of computing resources affects model performance. These laws include pretraining scaling, post-training scaling, and test-time scaling. These laws reflect how the AI field is evolving by applying additional computing technologies to increasingly complex AI use cases.

Sponsored

The recent rise of test-phase scaling has enabled new types of large language models (LLMs), such as AI reasoning models, to perform multiple inferences to handle complex problems while describing the steps required to solve the task. Test-phase scaling requires significant computing resources to support AI reasoning, further driving the demand for accelerated computing.

What is pre-training extension?

Pre-training scaling is an original law of AI development. It demonstrates that by increasing the size of the training dataset, the number of model parameters, and the amount of computing resources, developers can expect predictable improvements in model intelligence and accuracy.

Data, model size, and computation are all closely related. Research indicates that when larger models are fed more data, their overall performance improves. To achieve this, developers must scale up computation, which requires powerful accelerated computing resources to run larger training workloads.

This principle of pre-training scaling has enabled groundbreaking capabilities in large models and spurred significant innovation in model architectures, including the rise of transformer models with billions and trillions of parameters, mixtures of experts models, and new distributed training techniques, all of which require massive computational resources. Meanwhile, the relevance of the pre-training scaling principle continues to grow as humanity continues to generate more and more multimodal data—text, images, audio, video, and sensor information—that will be used to train future powerful AI models.

▲ Pre-training extension is a fundamental principle in AI development, linking the size of models, datasets, and computations to the effectiveness of AI. The hybrid expert model shown above is a commonly used model architecture for AI training.

What is Post-Training Scaling?

Pre-training large, basic models isn't suitable for everyone and requires significant investment, skilled experts, and datasets. However, once organizations have pre-trained and released models, they can allow others to use them as a foundation for their own applications, lowering the barrier to AI adoption.

This post-training process drives additional cumulative demand from businesses and the broader developer community for accelerated computation, and improves the specificity of models and their relevance to the use cases their organizations require. Post-training scaling primarily enhances the skills of a model for its intended work. For example, a large language model can be post-trained and scaled to handle tasks like sentiment analysis or translation, or to understand terminology in fields like healthcare or law.

The post-training scaling law posits that the performance of pre-trained models can be further improved in terms of computational efficiency, accuracy, or domain specificity using techniques such as fine-tuning, pruning, quantization, distillation, reinforcement learning, and synthetic data augmentation.

Reinforcement learning (RL) is a machine learning technique that uses reward models to train agents to make decisions tailored to specific use cases. The agent's goal is to make decisions that maximize cumulative rewards over time as it interacts with its environment. For example, a large chatbot language model might be positively reinforced by users' "like" responses. This technique is called reinforcement learning with human feedback (RLHF). Another newer technique, reinforcement learning with AI feedback (RLAIF), uses feedback from AI models to guide the learning process and streamline post-training tasks.

To support post-training expansion, developers can augment or supplement fine-tuning datasets with synthetic data. Supplementing real-world datasets with AI-generated data helps models improve their ability to handle edge cases that were underrepresented or omitted in the original training data.

Post-training extensions use techniques like fine-tuning, pruning, and distillation to refine pre-trained models for greater efficiency and task relevance.

What is a beta extension?

Extended test phase, also known as long thinking, occurs during the inference process. While traditional AI models quickly generate a one-time answer to a user prompt, models using this technology allocate additional computational work during the inference process, allowing the model to reason through multiple possible responses before arriving at the optimal answer.

This AI reasoning process can take minutes or even hours to generate complex custom code for developers, and difficult queries can require more than 100 times the amount of computing power compared to a single inference of a traditional large language model, because traditional large language models are unlikely to produce the correct answer to complex questions on the first try.

This test-phase computing capability allows AI models to explore different solutions to problems and break down complex requirements into multiple steps, in many cases showing users their work during reasoning. Research has found that when AI models are given open-ended prompts that require multiple reasoning and planning steps, the expanded test-phase approach can produce higher-quality responses.

There are various computational methods used during the testing phase, including "chain-of-thought prompts" that break down complex problems into a series of simpler steps, "majority sampling" that generates multiple responses to the same prompt and selects the most frequently occurring answer as the final output, and "search" processes that explore and evaluate multiple paths within a response tree. Furthermore, post-training extension methods similar to best-solution search sampling can be used to optimize long-term thinking during inference to optimize responses that align with human preferences or other goals.

▲Test phase expansion technology enhances AI reasoning capabilities by allocating additional computations, enabling the model to effectively solve complex multi-step problems

How to perform AI reasoning during the testing phase

The rise of computational technologies in the testing phase is enabling AI to provide informed, helpful, and increasingly accurate responses to complex, open-ended user queries. These capabilities are crucial for the detailed, multi-faceted reasoning tasks expected of autonomous agent-based AI and physical AI applications. They can provide users across various industries with powerful assistants to accelerate their work, thereby improving efficiency and productivity.

In healthcare, a model can use test-phase expansion techniques to analyze large amounts of data, infer the progression of a disease, and predict potential complications of new treatments based on the chemical structure of the drug molecule. Alternatively, it can comb through a database of clinical trial data to recommend treatment options that suit an individual's condition and share its reasoning about the pros and cons of different studies.

In retail and supply chain logistics, long-term thinking helps solve complex decisions required for both near-term operational challenges and long-term strategic objectives. Reasoning technology can simultaneously predict and evaluate multiple scenarios, helping companies mitigate risk and address scalability challenges. This enables more accurate demand forecasting, streamlined supply chain routing, and procurement decisions that align with organizational sustainability initiatives.

For global companies, the technology can be used to draw up detailed business plans, generate complex code to debug software, or optimize the routes of trucks, warehouse robots, and self-driving taxis.

AI reasoning models are developing rapidly. OpenAI o1-mini and o3-mini, DeepSeek R1, and Google DeepMind's Gemini 2.0 Flash Thinking were all released in the past few weeks, and more new models are expected to be released soon.

These models require a large amount of computing during the inference process to reason about complex problems and produce correct answers. This means that companies need to expand accelerated computing resources to provide next-generation AI inference tools that can solve complex problems, write code, and plan multi-step processes.

Tags: AINvidiaArtificial wisdom
ShareTweetShare
mashdigi news content

mashdigi news content

Leave a comment Cancel reply

Your e-mail address Will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

  • Popular
  • 留言
  • Latest
Sony unveils the Xperia PRO, a flagship phone under development, supporting both sub-6GHz and millimeter wave bands

Sony unveils the Xperia PRO, a flagship phone under development, supporting both sub-6GHz and millimeter wave bands

2020-02-24
Interview/Sony's new flagship phone, the Xperia 5, is positioned as a "Compact" phone?

Interview/Sony's new flagship phone, the Xperia 5, is positioned as a "Compact" phone?

2019-09-07
The Raspberry Pi Foundation announced the Raspberry Pi 5 development board, indicating that supply chain issues have improved.

The Raspberry Pi Foundation announced the Raspberry Pi 5 development board, indicating that supply chain issues have improved.

2023-09-28
Google will prioritize non-AMP pages in mobile search results starting next May

Google will prioritize non-AMP pages in mobile search results starting next May

2
Ubuntu's parent company confirms plans for IPO, but no specific timetable

Ubuntu's parent company confirms plans for IPO, but no specific timetable

1

Sony Interactive Entertainment confirms that PlayStation 5 is only backwards compatible with most PlayStation 4 games

0
NVIDIA further explains the details of the "Blackwell" display architecture, maintaining computing flexibility while achieving higher performance output.

In response to market rumors of "remote card locking," NVIDIA released a data center monitoring tool, reiterating that it has no hardware tracking or backdoor mechanisms.

2025-12-12
The Game Awards 2025: *Light and Shadow: Expedition 33* sweeps 9 awards, information on multiple new titles revealed.

The Game Awards 2025: *Light and Shadow: Expedition 33* sweeps 9 awards, information on multiple new titles revealed.

2025-12-12
Laura is back! Crystal Dynamics officially unveils its new title, Tomb Raider Catalyst, expected to launch across platforms in 2027.

Laura is back! Crystal Dynamics officially unveils its new title, Tomb Raider Catalyst, expected to launch across platforms in 2027.

2025-12-12
mashdigi-Technology, new products, interesting news, trends

Copyright © 2017 mashdigi.com

  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Follow us

Welcome back!

Login to your account below

Forgotten Password?

Retrieve your password

Hãy nhập tên người dùng hoặc địa chỉ email để mở mật khẩu

Log In
No Result
View All Result
  • About mashdigi.com
  • Place ads
  • Contact mashdigi.com

Copyright © 2017 mashdigi.com

Go to Mobile Version