| | | | Here is what we see. | A man in a black leather jacket stood on a stage in San Jose for two hours. Most coverage focused on the show — the confetti, the product names, the big numbers. But sit in the quiet after the noise fades, and a different picture comes into view. This was not a product launch. It was an infrastructure argument. The kind that enterprise teams file on Monday and act on for the next year and a half. Huang did not just announce chips. He laid out the plumbing of a new utility — and the effects are structural. |
| |
| | |
| | | | | The Factory Floor You Can't See | Start with the number that matters. NVIDIA said it will face a $1 trillion backlog in chip orders by year-end — double the estimate from twelve months ago. Annual revenue climbed from $27 billion in 2022 to $216 billion last year. Analysts now expect the company to pass $330 billion in the year ahead. These are not guesses on a whiteboard. They are purchase orders piling up at every major cloud provider on the planet. | But the real signal is not the size of the backlog. It is what the backlog is for. The center of gravity in AI has shifted — quietly — from training models to running them. Huang said it plainly at GTC 2026: "The inference inflection has arrived." Training taught the model to think. Inference is the moment it speaks, writes, creates, and acts. Inference never stops running. | Huang broke the math down clearly. A $50 billion AI factory built on NVIDIA's design produces tokens at roughly 10x the output of cheaper options. The price tag looks higher. The cost per token is lower. This is not a chip pricing story. It is a total-cost story — the kind that CFOs feel in their bones. | The hum you hear is not hype. It is cooling towers spinning up in data centers that now work less like server farms and more like factories — factories whose product is the token itself. | |
| |
| | |
| | | | | Disaggregated Inference and the Architecture Beneath | Here is where the first signal shows up for builders. NVIDIA did not just release a faster GPU. It unveiled Vera Rubin, a platform that blends CPUs, GPUs, networking, storage, and security into one rack-scale system — what Huang called "a computing structure for the era of AI agents." The company claims it can train large models with one-fourth the GPUs of the prior Blackwell generation. It also delivers up to 10x higher inference output per watt at one-tenth the cost per token. | The deeper move is disaggregated inference. Huang explained the idea on the All-In Podcast: break the inference pipeline into stages — prefill and decode — and route each stage to the best chip for the job. Vera Rubin handles the heavy prefill work. Groq-derived chips, gained through a multi-billion dollar licensing deal, tackle decode — the step that spits out the answer. NVIDIA evolved from a GPU company to an AI factory company. The factory floor now holds mixed silicon running under a single system called Dynamo. | That matters because rivals like Google's TPU and Amazon's Trainium are circling the inference market. NVIDIA's response is not to win on any single chip test. It is to own the full stack — every layer, every rack, every software layer — so that the switching cost becomes structural, not simple. Roughly 40% of NVIDIA's business already depends on buyers who need the whole factory, not just a chip. | | Sponsored by Brownstone Research | Jensen Huang's Shocking Announcement | NVIDIA's revolutionary new invention just solved the #1 chokepoint that's been strangling big AI companies. | And Tech legend Jeff Brown — the Silicon Valley insider who called NVIDIA before it skyrocketed more than 30,000%... | ... says a shocking announcement by NVIDIA CEO Jensen Huang could make a lot of early investors rich. | | |
| |
| | |
| | | | | The Agent Inflection and the $50 Trillion Quiet Bet | If inference is the new factory output, agents are the new workers on the floor. Huang named three waves that arrived in quick order: generative AI, reasoning AI, and now agentic AI. Each step grew compute demand by roughly 100x. In just two years, total compute needs jumped by a factor of 10,000x. That is not a forecast. It is a count of what already happened. | The agentic shift is not about chatbots getting smarter. It is about software that reads files, browses the web, runs code, and uses outside tools. NVIDIA announced NemoClaw, an enterprise-grade add-on to the open-source OpenClaw framework. It lets teams build governed, secured AI agents without complex setup. Huang called OpenClaw "the blueprint, the operating system of modern computing" — open source, running everywhere. Policy rules ensure agents never hold all three keys — data access, code execution, and outside links — at the same time. | Then there is the longer arc: physical AI. Huang framed it as tech's first chance to reach a $50 trillion market — robotics, self-driving cars, factory automation — that has had almost no software until now. NVIDIA announced deals with over 15 humanoid robotics firms and a Physical AI Data Factory Blueprint with Microsoft Azure and Nebius. Real-world data is scarce. Synthetic data plus simulation can turn compute into the raw material these systems need. This is a ten-year bet that is now, by Huang's count, a multi-billion dollar business nearing $10 billion a year. | | The Utility Thesis — And What You Do With It | Strip away the stage lights and the arena crowd. The lasting takeaway is structural. NVIDIA wants AI to stop being seen as a type of software and start being treated as a utility-scale infrastructure project — with NVIDIA's hardware and software at every layer. Tokens are the product. Inference is the factory process. Agents are the labor force. The physical world is the next market. | For leaders and investors, the mental model is simple. The companies that win the next cycle will not be the ones with the flashiest demo. They will be the ones that control the cost curve of token output — the plumbing, the wiring, the rails beneath every AI interaction. Huang put it bluntly: "Even when the chips are free, it's not cheap enough" if you cannot keep pace with the design. | Three signals to file: | - Inference costs now drive the buying choice. Total cost per token — not chip price — is the metric that matters from now to 2030. | - Agentic software needs a governance layer. Any team deploying agents without policy rules is building on sand. Watch the OpenClaw and NemoClaw ecosystem closely. | - Physical AI is no longer a demo. With $10 billion in yearly revenue and a growing robotics partner base, the timeline for viable physical AI has shortened fast. | The arena in San Jose has gone quiet. The confetti has been swept. But beneath the surface, the switchyard is humming — and the tokens are already flowing. This is plumbing, not hype. |
| |
| | |
| | | | |  | | | Brownstone Research | |
|
|
|
|
|
|
| |
| | |
| | |
|
Tidak ada komentar:
Posting Komentar