Nvidia said the revenue opportunity for its artificial intelligence chips could reach at least $1 trillion by 2027, as the company pivots more aggressively toward the rapidly expanding market for real-time AI computing.
The forecast was outlined by CEO Jensen Huang during the company’s annual Nvidia GTC developer conference in San Jose, California, where the chipmaker introduced new processors and system designs aimed at accelerating AI responses for large-scale applications.
“The inference inflection has arrived,” Huang said during the keynote address delivered in a packed arena with more than 18,000 attendees, highlighting how the AI industry is transitioning from building models to deploying them widely.
Register for Tekedia Mini-MBA edition 20 (June 8 – Sept 5, 2026).
Register for Tekedia AI in Business Masterclass.
Join Tekedia Capital Syndicate and co-invest in great global startups.
Register for Tekedia AI Lab.
The estimate represents a sharp increase from the roughly $500 billion AI infrastructure opportunity Nvidia previously projected for 2026, reinforcing investor expectations that global demand for AI computing capacity will continue expanding at a rapid pace.
Shares of Nvidia — currently the world’s most valuable publicly traded company with a market capitalization exceeding $4.3 trillion — briefly rose following the announcement before closing about 1.6% higher.
AI industry entering deployment phase
The company’s updated forecast indicates a broader shift taking place across the artificial intelligence industry. Over the past two years, the biggest technology companies have spent hundreds of billions of dollars acquiring computing hardware to train increasingly sophisticated AI models. Now the focus is rapidly moving toward inference — the process where trained models generate responses, make predictions, or execute tasks in real time for users.
This stage of AI computing is expected to be even larger in scale than the training phase because it involves serving millions or potentially billions of user queries daily across applications such as chatbots, search engines, productivity tools, and autonomous systems.
Companies including OpenAI and Anthropic are rapidly expanding AI services to support growing user bases, while technology firms such as Meta Platforms are integrating AI assistants into social media platforms used by billions of people. As a result, demand for hardware capable of delivering AI responses instantly — with minimal latency — is rising sharply.
While Nvidia dominates the market for chips used to train large language models, inference computing has become a more competitive arena. Large technology companies have begun designing their own specialized processors optimized for running AI models efficiently at scale.
Meta, for example, has invested heavily in developing custom AI chips for its internal infrastructure, while several cloud providers are also building proprietary silicon.
To maintain its leadership, Nvidia unveiled a new architecture designed specifically for high-performance inference workloads. The company’s upcoming Vera Rubin AI chip will perform a stage known as “prefill,” in which human prompts are converted into machine-readable tokens that AI systems can process.
The second stage of the process, called “decode,” will be accelerated by chips from startup Groq, which specializes in extremely fast AI inference.
Nvidia licensed Groq’s technology in a deal valued at $17 billion last year, denoting how critical inference performance has become for the next phase of AI deployment. The company aims to reduce the time it takes AI systems to generate responses — a key metric for applications ranging from digital assistants to automated customer service tools — by combining its own processors with Groq’s specialized hardware.
Even as competitors develop alternative AI chips, Nvidia continues to benefit from a powerful ecosystem built around its programming platform, CUDA. The CUDA software environment allows developers to design algorithms optimized for Nvidia hardware, creating a large installed base that reinforces the company’s dominance.
“The installed base is what attracts developers who then create the new algorithms that achieve the breakthrough technologies,” Huang said.
“We are in every cloud. We’re in every computer company. We serve just about every single industry.”
That ecosystem advantage has made it difficult for competing chip architectures to gain widespread adoption, even when they offer specialized performance improvements.
Next Generation Of AI Processors
Huang also introduced a new processor called the Feynman AI chip, named after the Nobel Prize-winning physicist Richard Feynman. The chip forms part of Nvidia’s long-term roadmap for advancing AI computing capabilities, with future processors expected to deliver significantly higher performance for both training and inference workloads.
The company is also investing heavily in technologies that enable faster communication between processors in large AI data centers. Analysts expect Nvidia to elaborate further on its recent $2 billion investments in optical networking companies Lumentum and Coherent Corp., which manufacture laser-based components used to transmit data between chips using beams of light.
Optical interconnects are increasingly essential as AI systems grow to include tens of thousands of processors operating simultaneously within massive data center clusters.
Beyond technology companies, national governments are also becoming major buyers of AI infrastructure. Countries, including Saudi Arabia, are investing heavily in sovereign AI systems designed to support national data processing, research, and digital services. These projects often rely on Nvidia’s processors, reinforcing the company’s role as a foundational supplier to the global AI ecosystem.
At the same time, AI hardware has become a key element of technological competition between the United States and China, with export controls limiting the sale of some advanced chips to Chinese companies. Despite those geopolitical tensions, Nvidia continues to release open-source AI software tools, positioning itself as a central platform provider for developers worldwide.
Nvidia’s $1 trillion revenue opportunity forecast highlights how the AI infrastructure boom is expected to extend well beyond the initial wave of model development.
As artificial intelligence becomes embedded across industries — from healthcare and finance to logistics and manufacturing — the amount of computing power required to support those systems will expand dramatically. Analysts say the next stage of growth will be driven by applications that require real-time AI responses, including autonomous vehicles, intelligent robotics, and advanced digital assistants.
If those technologies scale globally, demand for AI computing infrastructure could grow far beyond current projections. The transition from model training to real-time inference is being interpreted as the next major phase of growth, especially for Nvidia, which has already become the most valuable company in the world, largely due to the AI boom.



