Nvidia didn’t simply catch the AI wave; it successfully turned the tide.
Over the previous couple of years, the favored chipmaker has advanced from a tech stalwart to a market superpower, surpassing $4 trillion in market capitalization.
The inventory has now became the S&P 500’s engine, using a once-in-a-generation build-out of AI knowledge facilities.
Beneath the hood, Nvidia owns the excessive finish of AI compute. Estimates present that it instructions an 80% to 90% share of AI accelerators. Everybody from OpenAI to the most important hyperscalers is scrambling to purchase its chips.
Income has adopted, with AI orders pushing quarterly gross sales to information, whereas software program and networking have solely deepened the moat.
For perspective, Nvidia’s gross sales within the pre-ChatGPT period (fiscal 2022) have been at $26.91 billion and have now surged to $130.5 billion in fiscal 2025 (a 385% soar).
Now comes the subsequent pivot: Vera Rubin, its brand-new CPU+GPU platform, tailored for powering the way forward for AI as we all know it.
With the platform, Nvidia isn’t simply refreshing chips; it’s extending the cadence whereas permitting hyperscalers to plan years upfront.
Vera Rubin may mark the beginning of “AI Wave 2.0,” extending Nvidia’s lead within the data-center race.
Picture supply: Morris/Bloomberg through Getty Pictures
What’s Nvidia’s “Vera Rubin” platform?
Nvidia’s subsequent main platform isn’t only a fancy new nickname, but it surely’s maybe the subsequent AI development cycle. “Vera Rubin” effectively layers a brand new Arm-based Vera CPU with the Rubin GPU, a direct successor to the present Blackwell chips, designed to deal with the big AI inference and coaching wants.
The flagship rack, the Vera Rubin NVL144 CPX, is anticipated to ship almost8 exaflops of AI efficiency together with 100 terabytes of quick reminiscence, which is about 7.5 occasions extra efficiency than the present GB300 NVL72 system.
One other main edge for Nvidia is that the brand new platform is tuned for dealing with next-generation workloads, together with issues like million-token coding, generative video, and autonomous brokers.
Associated: Former Intel CEO drops curt 2-word verdict on AI
On the {hardware} facet, Rubin considerably improves compute density (as much as 50-100 PFLOPs FP4 per GPU), reminiscence bandwidth (with HBM4 and GDDR7 variants), and networking by means of NVLink 144 and Spectrum X.
Additionally, Nvidia says the platform has already been taped out and is at the moment “in fab” at TSMC, with basic availability anticipated in late 2026.
Might Vera Rubin spark one other AI wave?
Many AI specialists agree that the business’s subsequent main leap will not come from smarter algorithms, however from quicker, denser infrastructure, and Rubin is correct in that profitable wheelhouse.
With its efficient new rack-scale design and economics, analysts say Rubin may probably unlock the subsequent wave of AI adoption throughout hyperscalers and enterprises.
Right here’s why Vera Rubin issues
- Greater context home windows: With a whopping 100 TB of rack reminiscence and 1.7 PB/s bandwidth, Rubin is ready to deal with colossal datasets and prompts on-rack, reducing latency and chatter.
- Higher ROI math: Nvidia claims prospects can successfully monetize almost $5 billion in token income per $100 million invested in Rubin infrastructure. If that declare is remotely correct, that considerably slashes inference prices whereas making always-on copilots commercially viable.
- Predictable rollout: Nvidia’s regular cadence de-risks hyperscaler planning the place consumers can simply deploy Blackwell now and slot Rubin in 2026 with out retooling their stacks.
- Greater pricing energy: With HBM4, NVLink 144, Spectrum X networking, together with full rack-scale integration, Vera Rubin is more likely to carry considerably increased ASPs than Blackwell, increasing the corporate’s gross margins within the course of.
Nvidia’s GPU evolution at a look
- Celsius/ Kelvin/Rankine/Curie (1999-2004): A large step-up from primarily fixed-function graphics to really programmable shaders, opening the door for basic compute.
- Tesla (2006): The CUDA period begins, laying the inspiration for a unified programming mannequin that allowed the operating of non-graphics math on GPUs.
- Fermi (2010): Beefed up double-precision and higher reminiscence hierarchy.
- Kepler (2012): Higher effectivity and extra parallelism per watt, facilitating huge energy financial savings for knowledge facilities and in scaling core counts.
- Maxwell (2014): Main perf/watt leap and smarter caching, enabling stronger real-world throughput.
- Pascal (2016): P100 with HBM2 and NVLink confirmed up, which led to superior reminiscence bandwidth and faster GPU-to-GPU hyperlinks, facilitating the primary really giant AI coaching clusters.
- Volta (2017): First Tensor Cores purpose-built for deep studying, delivering unbelievable speedups for coaching neural networks.
- Turing (2018): Tensor + RT Cores for real-time ray tracing on desktops.
- Ampere (2020): Wider rollout of Tensor Cores, together with extra strong effectivity throughout knowledge heart and shopper.
- Hopper (2022): Transformer Engine optimized for LLMs pushing generative AI into overdrive.
- Blackwell (2025): Colossal coaching/inference soar backed by various reminiscence choices and quicker interconnects.
- Vera Rubin (2026, deliberate): Vera CPU + Rubin GPU as an built-in, rack-scale AI platform custom-built to deal with ultra-long context home windows, heavier reminiscence footprints, and agentic/video workloads.
AMD’s reply to Nvidia’s Vera Rubin: the way it stacks up
AMD’s catching as much as Nvidia’s Vera Rubin shortly.
Its highly effective Intuition line has moved swiftly from right this moment’s MI300X to MI350, then MI450 in 2026, with Oracle planning to deploy 50,000 MI450s beginning Q3 2026, a significant nod of approval.
AMD additionally introduced Helios, which is a robust rack-level system, probably becoming in almost 50% extra reminiscence per rack than Nvidia’s comparable Vera Rubin setup.
Associated: Jamie Dimon drops shocking tackle AI shares
Nvidia’s Rubin is betting on a break up design, focusing extra on a compute-heavy configuration that facilitates uncooked pace, in addition to one other tuned for extra demanding long-context duties. AMD’s pitch, although, is easier, because it goals to pack extra reminiscence per GPU and open-source software program (ROCm) to lure builders away from CUDA.
Finally, nonetheless, the winner comes all the way down to throughput per greenback and software program ease, the place Nvidia nonetheless has an edge.
Wall Avenue bets on Nvidia’s subsequent leap, the Rubin chip cycle
Wall Avenue is singing the praises of Nvidia’s GPU platform rhythm.
Goldman Sachs not too long ago reiterated a stable purchase ranking, modeling 2026 EPS to be 10% above consensus, on the again of the Rubin ramp and resilient hyperscaler spending. Morgan Stanley additionally raised its value goal on the inventory to $200 on a 33× 2026 p/e, pointing to easing provide bottlenecks.
Citi went even increased at $210, flagging Rubin CPX as a brand new development driver whereas lifting its AI infrastructure forecast to $490 billion by 2026. On the identical time, HSBC’s Frank Lee has a brand new Avenue-high ranking at $320, arguing that mega-deals just like the one with OpenAI may double Nvidia’s data-center income by FY27.
It’s important to notice that Nvidia’s Vera Rubin combo arrives at a vital juncture.
At GTC 2025, CEO Jensen Huang pitched the subsequent part as agentic AI, and Rubin is tailored for dealing with richer world fashions and video.
Associated: Veteran dealer who purchased Nvidia at $13 resets inventory value goal
