New Technology / Robotics
Track robotics trends, industrial automation, machine intelligence and commercial deployment signals through curated technology summaries.
【直播|AI即時中字翻譯】輝達 NVIDIA GTC 2026 大會 #黃仁勳 主題演講
Topic
Advancements in AI and Computing Infrastructure
Key insights
- AI has reached a critical inflection point, enabling productive work and marking the inference inflection
- Computing demand surged by 10,000 times in two years, driven by AIs need for reasoning and token generation
- NVIDIAs infrastructure offers the lowest cost for AI, ensuring long-term utility and reduced user costs
- NVIDIAs investments in AI infrastructure enhance capabilities across all phases, especially in inference
- Collaboration with companies like Anthropic and MSL highlights NVIDIAs dominance, accounting for one third of global AI compute
- Open source AI models are nearing the frontier, indicating widespread adoption across various domains
Perspectives
NVIDIA's advancements in AI and computing infrastructure are significant, but skepticism remains regarding the scalability and universal applicability of these technologies.
NVIDIA's Vision and Achievements
- Highlights AIs critical inflection point enabling productive work
- Claims computing demand surged by 10,000 times in two years
- Proposes NVIDIAs architecture as the lowest cost and most reliable for AI
- Demonstrates significant performance improvements in AI infrastructure
- Announces the operational success of the Vera Rubin architecture
- Describes the integration of GROC technology to enhance performance
Skepticism and Challenges
- Questions the metrics behind the claimed 10,000 times increase in computing demand
- Challenges the assertion of NVIDIAs architecture being universally cost-effective
- Raises concerns about the representativeness of synthetic data for training AI
- Critiques the assumption that all enterprises can adopt OpenClaw seamlessly
- Highlights potential integration hurdles in transitioning to new AI models
- Questions the scalability of performance gains across diverse workloads
Neutral / Shared
- Acknowledges the rapid advancements in AI technology and infrastructure
- Recognizes the growing importance of AI in various industries
- Mentions the collaborative efforts in developing AI models and systems
Metrics
revenue
$1 trillion USD
projected revenue through 2027
This projection suggests a strong market confidence in AI's future profitability.
I see through 2027, at least $1 trillion.
revenue
60%
percentage of revenue from hyperscalers
This indicates a strong reliance on large-scale AI operations for revenue generation.
60% of our business is hyperscalers.
performance
700 tokens per second tokens/second
token speed before software update
This baseline highlights the dramatic improvement achieved through optimization.
before Nvidia updated everything, and all of our algorithms, and software, and all the technology that we bring to bear, about 700 tokens per second
throughput
35 times more throughput per megawatt times
performance improvement of Grok3 LPX Rack
This metric indicates a significant enhancement in energy efficiency for AI processing.
35 times more throughput per megawatt
compute_power
40 million times more compute in just 10 years times
growth in compute power of Vera Rubin
This illustrates the exponential growth in AI capabilities over a decade.
40 million times more compute in just 10 years
performance
twice the performance per watt x
comparison with current CPUs
This highlights NVIDIA's leadership in energy-efficient computing.
the Vera system, twice the performance per watt of any CPUs in the world today.
business_projection
multi-billion dollar business USD
projected revenue from standalone CPUs
Indicates significant market potential for NVIDIA's CPU offerings.
This is already for sure going to be a multi-billion dollar business for us.
gpu_connections
144 GPUs units
number of GPUs connected in Kiber rack
Demonstrates the scalability of NVIDIA's compute architecture.
enables us to connect 144 GPUs in one MV link domain.
Key entities
Timeline highlights
00:00–05:00
AI has reached a critical inflection point, enabling productive work and marking the inference inflection. Computing demand surged by 10,000 times in two years, driven by AI's need for reasoning and token generation.
- AI has reached a critical inflection point, enabling productive work and marking the inference inflection
- Computing demand surged by 10,000 times in two years, driven by AIs need for reasoning and token generation
- NVIDIAs infrastructure offers the lowest cost for AI, ensuring long-term utility and reduced user costs
- NVIDIAs investments in AI infrastructure enhance capabilities across all phases, especially in inference
- Collaboration with companies like Anthropic and MSL highlights NVIDIAs dominance, accounting for one third of global AI compute
- Open source AI models are nearing the frontier, indicating widespread adoption across various domains
05:00–10:00
NVIDIA's architecture is designed to support a wide range of AI applications, establishing it as a reliable and cost-effective platform for AI infrastructure. The company's revenue is significantly driven by hyperscalers, with 60% of its business stemming from this sector, indicating strong demand for advanced AI capabilities.
- NVIDIAs architecture supports diverse AI applications, making it the lowest cost and most reliable platform for AI infrastructure
- Investing in NVIDIAs infrastructure ensures performance and cost-effectiveness, crucial for the trillion-dollar AI market
- 60% of NVIDIAs revenue comes from hyperscalers, reflecting strong demand for AI in large-scale operations
- Hyperscaler workloads are shifting towards deep learning, increasing reliance on NVIDIA GPUs for advanced AI tasks
- NVIDIAs ecosystem of AI partners accelerates compute resource consumption, reinforcing its market position
- Technological advancements like MV link 72 and MV FP4 enhance performance and energy efficiency
10:00–15:00
NVIDIA has achieved a significant performance improvement in AI infrastructure, with advancements yielding 35 to 50 times higher performance per watt. This positions the company favorably in a power-constrained environment where maximizing token production is critical.
- NVIDIAs advancements have achieved 35 to 50 times higher performance per watt, establishing a competitive edge in AI infrastructure
15:00–20:00
Nvidia's software update significantly increased token speeds from 700 to nearly 5,000, highlighting the importance of optimized architecture in data centers. The evolution of AI infrastructure, including the introduction of the Vera Rubin architecture, positions companies to enhance their token generation capabilities.
- Nvidias software update boosted token speeds from 700 to nearly 5,000, demonstrating extreme co-designs impact
- Data centers are transforming into token generation factories, necessitating optimized architecture
- Every AI company will focus on token factory efficiency as tokens become the new revenue-driving commodity
- The DGX1 launch in 2016 initiated a new deep learning era tailored for AI researchers
- Hopper introduced the first GPU with the FPA transformer engine, marking the start of the Generative AI era
- Blackwell achieved 130 terabytes per second bandwidth with 72 GPUs via NVLink, redefining AI supercomputing
20:00–25:00
The Grok3 LPX Rack, integrated with Vera Rubin, significantly enhances AI efficiency with 35 times more throughput per megawatt. Vera Rubin achieves 40 million times more compute power in a decade, underscoring rapid advancements in AI technology.
- The Grok3 LPX Rack, integrated with Vera Rubin, boosts throughput by 35 times per megawatt, enhancing agentic AI efficiency
- Vera Rubin, a revolutionary AI supercomputer, achieves 40 million times more compute power in a decade, highlighting rapid AI evolution
- Designed for large language models, the Vera Rubin system emphasizes fast memory access, crucial for demanding AI workloads
- A new CPU optimized for high single-threaded performance and energy efficiency utilizes LPDDR5 technology, enhancing data processing
- Vera Rubins 100% liquid cooling reduces installation time from two days to two hours, showcasing supercomputer design advancements
- Operating at 45 degrees Celsius, the cooling system lowers energy costs and reallocates resources to computational tasks
25:00–30:00
NVIDIA's Vera system offers twice the performance per watt compared to current CPUs, reinforcing its position in energy-efficient computing. The company is now successfully selling standalone CPUs, which are expected to become a multi-billion dollar business.
- NVIDIAs Vera system delivers twice the performance per watt of current CPUs, solidifying its leadership in energy-efficient computing
- The company is now selling standalone CPUs, projected to become a multi-billion dollar business
- The CX9, powered by the Vera CPU, marks a significant leap in data center CPU technology
- NVIDIAs Bluefield 4 STX enhances data handling and storage efficiency with its new storage platform
- Optimized MV link rack technology simplifies data center cabling, streamlining infrastructure
- Ruben Ultra compute nodes can connect 144 GPUs, boosting computational power and scalability