View: 14|Reply: 1

FWD: NVIDIA’s Vera Rubin — The Beginning of AI as Infrastructure

[Copy link]

1

threads

2

posts

15

credits

Novice

Rank: 1

credits
15
Published in yesterday 23:58 | Show all floors |Read mode
At CES 2026, NVIDIA CEO Jensen Huang delivered one of the most consequential keynotes in the company’s history. What stood out was not what was launched—but what wasn’t. There was no new consumer GPU, no RTX announcement, and no performance charts aimed at gamers. Instead, NVIDIA introduced Vera Rubin, a next-generation AI supercomputing platform that reframes how artificial intelligence infrastructure is designed, deployed, and scaled.
The message from the stage was clear: AI’s bottleneck is no longer individual chip performance, but system-level efficiency, cost, and scalability. Rubin is NVIDIA’s answer to that challenge.
Rubin Is Not a GPU Launch — It’s a Platform Transition
Unlike previous NVIDIA generations—Hopper, Blackwell, or earlier architectures—Rubin is not defined by a single chip. Instead, it is a rack-scale AI computing platform, built from the ground up as a unified system.
NVIDIA positions Rubin as an AI supercomputer architecture, not a component upgrade. The platform integrates six purpose-built technologies that are co-designed to operate as a single AI engine:
  • Rubin GPU
    The core accelerator, featuring a new generation Transformer Engine optimized for large-scale inference and training workloads.
  • Vera CPU
    A new CPU designed specifically for AI reasoning and data orchestration, tightly coupled to the GPU via high-bandwidth links.
  • NVLink 6 Switch
    NVIDIA’s latest interconnect technology, enabling massive GPU-to-GPU bandwidth and near-linear scaling across racks.
  • ConnectX-9 SuperNIC
    High-performance networking optimized for AI clusters and low-latency communication.
  • BlueField-4 DPU
    A data processing unit that offloads networking, storage, and security tasks, while enabling AI-native memory and context management.
  • Spectrum-6 Ethernet Switch
    High-capacity Ethernet switching for hyperscale AI environments.

This six-component design reflects a fundamental shift: AI performance is now determined by how well compute, memory, networking, and security work together, not by raw GPU throughput alone.
From “Stacking GPUs” to Building AI Factories
In his CES keynote, Jensen Huang emphasized that the industry can no longer rely on simply adding more GPUs to solve AI’s scaling problems. Model sizes are growing into the hundreds of billions and trillions of parameters, and inference workloads increasingly involve long-context reasoning, agentic AI, and persistent memory.
Rubin addresses these challenges by treating the entire data center as a single AI computer.
The Rubin NVL72 System
One of the most striking demonstrations at CES was the Rubin NVL72 rack:
  • 72 Rubin GPUs
  • 36 Vera CPUs
  • Fully interconnected via NVLink 6
  • Aggregate bandwidth of up to 260 TB/s
  • Designed to behave like one massive logical GPU

This level of integration allows NVIDIA to dramatically reduce communication overhead, one of the largest inefficiencies in large-scale AI training and inference.
Rather than GPUs waiting on data, or CPUs idling while accelerators compute, Rubin coordinates the entire system so that compute, memory, and data movement remain continuously active.
Performance Gains That Redefine Economics
NVIDIA’s official figures—presented both on stage and in its press materials—highlight why Rubin represents more than an incremental upgrade.
Compared to Blackwell, Rubin Delivers:
  • Up to 5× improvement in inference performance
  • Up to 3.5× improvement in training performance
  • Up to 10× reduction in inference cost per token
  • Up to 4× reduction in the number of GPUs required for large MoE models

These gains are not achieved through brute-force compute alone. Instead, they result from:
  • Improved interconnect bandwidth via NVLink 6
  • Better CPU–GPU task coordination
  • Offloading of networking and storage overhead to DPUs
  • AI-native memory hierarchies designed for KV cache and long-context inference

The result is a platform that lowers the cost barrier for deploying large-scale AI, making advanced reasoning models economically viable for more organizations.
Implications for the Secondhand Hardware MarketSystems Become the Primary Asset
As AI infrastructure becomes more integrated, secondary markets can no longer treat GPUs as standalone commodities. Server configurations, interconnects, DPUs, and memory bandwidth increasingly determine real-world performance and resale value. Pricing and demand will reflect system context, not just accelerator model.
Earlier Generations Enter a New Phase
With Rubin entering production and broader deployment expected in the second half of 2026, platforms such as Blackwell and Hopper will gradually shift from frontline roles to secondary and niche use cases. These systems are likely to remain viable for cost-sensitive inference, hybrid deployments, and research workloads, creating renewed activity in secondary markets as assets are redeployed rather than retired.
Clear Timelines Enable Market Planning
NVIDIA’s stated deployment plans—beginning with major cloud providers in late 2026—provide rare visibility into the next infrastructure cycle. This predictability allows enterprises and data center operators to plan upgrades and divestments more deliberately, aligning primary adoption with secondary market supply.
A Broader Transition
Rubin signals that AI hardware is entering a more industrialized phase, where integration, efficiency, and lifecycle management matter as much as raw compute. In this environment, the ability to strategically sell GPUs and systems becomes part of infrastructure planning, not merely an end-of-life consideration. The shift from component-centric to system-centric AI infrastructure will shape both primary deployments and secondary market dynamics for years to come.
A Defining Moment for AI Infrastructure
CES 2026 made one thing unmistakably clear: the era of AI as a collection of GPUs is over.
With Vera Rubin, NVIDIA has drawn a line between past and future—between experimental scale-up and industrial-scale AI deployment. By integrating compute, memory, networking, storage, and security into a single platform, Rubin transforms AI from a resource-intensive experiment into a scalable, cost-controlled infrastructure.
The implications will ripple across cloud computing, enterprise IT, data center design, and global hardware markets for years to come.
AI is no longer just about faster chips. It’s about building systems that think at scale.
The article was originally published here.

1

threads

2

posts

15

credits

Novice

Rank: 1

credits
15
 Author| Published in 4 hourbefore | Show all floors
10 Best Places to Sell Your GPU for the Most Returns

If you’re looking to maximize cash returns from selling a used GPU, there are multiple effective channels — each with its own balance of price, effort, and safety. The best choice depends on whether you prioritize speed, convenience, or highest profit.

1. ITAD Companies (like BuySellRam.com)
Professional IT asset disposition firms provide quick, hassle-free offers, often with shipping support and fast payouts — ideal for businesses with multiple GPUs or enterprise hardware.

2. eBay
A large marketplace with broad buyer reach. It can yield high prices, especially for in-demand or rare cards, but seller fees and shipping logistics reduce net returns.

3. Facebook Marketplace
Local selling lets you avoid fees and shipping. You can meet buyers in person and get cash directly, but you must take precautions around safety and scams.

4. Reddit (r/Hardwareswap / Hardware communities)
Hardware-enthusiast communities often attract knowledgeable buyers willing to pay competitive prices — no intermediary fees, though you handle payment and shipping yourself.

5. Specialized Marketplaces (Swappa, tech forums)
Platforms like Swappa focus on electronics and often have lower fees than general marketplaces while still reaching a targeted audience.

6. Local Classifieds (Craigslist, community boards)
Good for quick, local cash sales. Like Facebook Marketplace, you avoid fees, but buyer quality and safety are variable.

7. Trade-In or Retail Programs
Trade-in options from stores (e.g., electronics retailers) offer convenience and sometimes instant store credit or cash, though often at lower payout levels.

8. Auction Sites / Collector Markets
For rare or vintage GPUs, auctions or niche events can attract premium bids — though outcomes are unpredictable and may involve auction fees.

9. Local Computer Shops / Resellers
Some local tech shops buy used GPUs either outright or on consignment, providing quick cash or trade credit with minimal hassle.

10. Friends, Family, Enthusiast Networks
Selling within your own network can be the safest and simplest option, often avoiding fees and shipping — if there’s a buyer willing to pay a fair price.

Key takeaways:
✔ Market reach vs. effort: Wider platforms often get higher prices but require more work or fees.
✔ Local options: Quick and fee-free but may offer lower average prices.
✔ Professional buyers: ITAD companies provide convenience, fast processing, and reliable payouts — especially good for bulk or business sellers.
You need to log in before you can reply login | Register

Points Rule

Quick reply Top Return list