At GTC 2025, NVIDIA CEO Jensen Huang emphasized that Blackwell architecture-based accelerator products are now in full production and expected to enter the market in the second half of this year. He also reiterated that computing power remains crucial in the development of artificial intelligence technology, and NVIDIA will continue to adhere to its previously proposed One Year Rhythm technology growth target. Therefore, NVIDIA not only plans to launch the next Rubin display architecture product in 2026, but also plans to advance the Feynman display architecture named after the renowned American physicist Richard Feynman in 2028. ▲At GTC 2025, NVIDIA reiterated that computing power remains the "truth" supporting the development of artificial intelligence technology. Blackwell architecture products are now in full production. Although previous accelerator products designed with the Blackwell architecture experienced delays due to design flaws, after resolving the issues in collaboration with TSMC, Jensen Huang stated that Blackwell architecture products are now in full production. NVIDIA will collaborate with numerous companies to launch various server application products, as well as with cloud providers. Blackwell architecture products will also be applied in telecommunications networks, edge computing, and even the autonomous vehicle and robotics markets, thereby accelerating the development of more artificial intelligence technology applications. ▲GPU accelerated designs are already widely used in many fields. ▲Blackwell display architecture products are now in full production and will be available for market deployment as early as the end of this year. In addition to emphasizing the full production of Blackwell architecture products, Huang also explained that even though artificial intelligence technology accelerates computational efficiency, the fundamental nature of computation still requires computing power, meaning that the underlying computing technology stacking remains necessary. Jensen Huang used the example of using a large-scale natural language processing (NLP) model to infer seating arrangements for different guests at a wedding banquet. While most current NLP models can arrive at an answer with a small number of words, the results may not meet requirements or may even be incorrect. To enable a large-scale NLP model to optimally adjust seating arrangements based on different guest relationships and needs, the inference process inevitably requires more words for deeper reasoning, significantly increasing the number of computations. Therefore, if faster processing speed and response time are desired, more computing power must be added, rather than relying solely on artificial intelligence computation. ▲Taking the seating arrangement of guests at a wedding banquet as an example, if artificial intelligence is to perform deliberate reasoning and calculation, the process must generate more than 20 times the amount of vocabulary and require more than 150 times the computing power. This stacking of computing power remains necessary and will continue to expand. In further explanation, Huang Renxun stated that the key aspects of artificial intelligence development include "cognition" and "inference." The former involves knowing and understanding the "acquired" information, while the latter analyzes and thinks to arrive at a reasonable answer. Current artificial intelligence technology, after converting user commands into vocabulary input, continuously generates more vocabulary during execution. These vocabulary words are then used as input in subsequent inference processes, and the most suitable answer is obtained through multiple iterations of inference. This process means that for artificial intelligence to arrive at a reasonable inference answer through "deliberation," its computational process must handle a much larger number of vocabulary words. To handle a large number of vocabulary words, higher computing power resources must be consumed. Furthermore, if artificial intelligence is to arrive at an answer faster, computing power must be further stacked. Even though NVIDIA has proposed NVIDIA Dynamo, an open-source inference software that can accelerate AI computing and reduce the overall cost of AI computing, in the long run, the computing power demand behind AI technology will still grow exponentially, and there may even be a much larger demand for computing power. ▲The open-source inference software NVIDIA Dynamo proposed this time mainly optimizes existing artificial intelligence computing, but the actual improvement is still limited, mainly relying on subsequent computing power stacking. ▲According to NVIDIA's approach, the operation of robots can be accelerated by different artificial intelligence operation methods to improve the smoothness of their work execution and improve the accuracy of their operation judgment. For example, the open-source Issac GR00T N1 allows robots to intuitively react to actions through an intuitive computing system, and performs more complete task inference through another deliberate system. This allows robots to improve the smoothness of their actions and make more accurate work judgments. Therefore, in his keynote speech at GTC 2025, Jensen Huang said that although the accelerator design of the Hopper display architecture has only been around for a few years, there is not much to say about it at present, given the current trend of computing power growth. The next development will be from the current Blackwell to the next Rubin, and will soon enter the next generation of Feynman display architecture. If we use the computing power performance of the Hopper display architecture as a benchmark, the Blackwell display architecture achieves approximately 68 times the computing power, while the subsequent Rubin architecture shows a growth rate of up to 900 times. Furthermore, if we use Hopper as a benchmark for cost per computation, we find that Blackwell requires only 0.13 times the cost of Hopper to achieve the same performance, and Rubin requires only 0.03 times the cost of Hopper. This means that with the same overhead, Blackwell and Rubin can drive significantly higher computing power.