Nvidia CEO Jensen Huang delivered a striking message for engineering talent during an appearance on the “All-In Podcast” episode published Thursday: top engineers who fail to consume hundreds of thousands of dollars worth of AI tokens annually are a cause for serious concern.
Huang stated he would be “deeply alarmed” if one of Nvidia’s $500,000-a-year engineers spent less than half that amount — $250,000 — on AI tokens over the course of a year.
“That $500,000 engineer at the end of the year, I’m going to ask them how much did you spend in tokens? If that person said $5,000, I will go ape something else,” he said.
Register for Tekedia Mini-MBA edition 20 (June 8 – Sept 5, 2026).
Register for Tekedia AI in Business Masterclass.
Join Tekedia Capital Syndicate and co-invest in great global startups.
Register for Tekedia AI Lab.
When asked whether Nvidia itself is spending $2 billion annually on tokens for its engineering team, Huang replied: “We’re trying to.”
He drew a direct analogy to outdated methods: “This is no different than one of our chip designers who says, ‘Guess what? I’m just going to use paper and pencil.’”
Huang argued that engineers who underutilize AI tokens are effectively limiting their own productivity and impact.
Tokens as a Recruiting and Productivity Tool
Huang went further, revealing that AI token budgets are already becoming a competitive recruiting lever in Silicon Valley.
“They’re going to make a few hundred thousand dollars a year, their base pay,” he said of engineers. “I’m going to give them probably half of that on top of it as tokens so that they could be amplified 10X.”
“It is now one of the recruiting tools in Silicon Valley: How many tokens comes along with my job?” Huang added. “And the reason for that is very clear, because every engineer that has access to tokens will be more productive.”
Tokens, the basic unit used by large language models to process and generate text, are typically charged on a per-thousand or per-million basis by providers such as OpenAI, Anthropic, Google, and others. Heavy usage can quickly become expensive, especially for engineers running large-scale experiments, fine-tuning models, or building complex agentic workflows.
Tokens as the “Fourth Component” of Compensation
Huang is not alone in viewing generous AI compute access as a critical talent differentiator. Business Insider reported earlier in March 2026 that tech companies are experimenting with offering token budgets alongside traditional salary, bonuses, and equity — effectively treating inference power as a new form of compensation.
Tomasz Tunguz of Theory Ventures described tokens as a potential “fourth component” of pay packages. Peter Gostev, AI capability lead at Arena (a startup focused on model performance benchmarking), suggested that frontier labs like OpenAI and Anthropic could create recruitment marketplaces listing token budgets alongside salary ranges.
Thibault Sottiaux, an engineering lead on OpenAI’s Codex team, noted on X that candidates increasingly ask how much compute they will receive.
Even OpenAI CEO Sam Altman has speculated about a future where compute access replaces traditional income support. In a May 2024 appearance on the same “All-In Podcast,” Altman mused: “I wonder if the future looks something more like Universal Basic Compute than Universal Basic Income, and everybody gets a slice of GPT-7’s compute. And they can use it, they can resell it, they can donate it to somebody to use for cancer research, but what you get is not dollars but this like slice, you own part of the productivity.”
As Engineers Shift to Token
Huang’s comments reflect Nvidia’s unique position at the center of the AI boom. As the dominant supplier of GPUs for training and inference, Nvidia benefits directly from skyrocketing token consumption across the industry. By framing heavy token usage as a productivity imperative — and even a recruiting tool — Huang is reinforcing the narrative that access to advanced AI compute is now a core requirement for top engineering talent.
The remarks also highlight a shift in how companies measure engineering productivity. Traditional metrics (lines of code, features shipped) are giving way to compute-intensive workflows: model experimentation, agent orchestration, large-scale data processing, and real-time inference. Engineers who underuse tokens may be seen as operating below their potential in an AI-native environment.
For talent acquisition, token budgets could become a powerful differentiator — especially as frontier models grow more expensive to run at scale. Startups and large tech firms alike may increasingly compete not just on salary and equity, but on how much high-quality inference capacity they can provide.
Huang’s stance is likely to accelerate the trend toward compute-inclusive compensation packages across Silicon Valley and beyond. As AI agents and multimodal models become central to software development, engineering roles will demand ever-larger token allocations — turning inference spend into a visible line item in hiring negotiations.
Nvidia itself stands to benefit disproportionately: more engineers consuming more tokens means more demand for Nvidia GPUs and cloud capacity. The company’s push to make heavy token usage a performance expectation — and even a hiring criterion — further cements its central role in the AI talent and productivity ecosystem.
The notion also underscores a broader philosophical shift: in the AI era, raw human intelligence is increasingly amplified (and measured) by access to compute. Engineers who maximize their token spend aren’t just more productive — they’re demonstrating mastery of the new tools defining the profession. For top talent, the question may soon be less “how much equity?” and more “how many tokens come with the job?”



