🦾 The Compute Stack Behind Physical AI: Inside NVIDIA’s System Architecture

A Newsletter for Entrepreneurs, Investors, and Computing Geeks

Happy Monday! Here’s what’s inside this week’s newsletter:

  • Deep dive: NVIDIA’s three-computer system for Physical AI, showing how dedicated training, simulation, and runtime platforms combine hardware, system software, and models to build and deploy real-world robotic intelligence.

  • Spotlights: Firgun Ventures launches a $250M fund focused on early growth-stage quantum companies, and a deep dive on the energy economics of inference.

  • Headlines: Semiconductor updates from SK hynix, TSMC, and imec, quantum networking and national initiatives, progress in photonic and neuromorphic hardware, and major AI data center and cloud projects.

  • Readings: Open chiplet and lifecycle approaches in semiconductors, quantum insights and investment outlooks, nonlinear photonic architectures, and neuromorphic systems for sensing and time synchronization.

  • Funding news: A quieter week concentrated in early-stage activity, with most financings in photonics and additional rounds in semiconductors and quantum, reflecting a pause in larger late-stage deals.

  • Bonus: Google’s accelerated TPU commercialization shows how diversified buyers and large-scale deployments could meaningfully challenge Nvidia’s dominance in AI compute.

Deep Dive: The Compute Stack Behind Physical AI – Inside NVIDIA’s System Architecture

Why Physical AI Is a Computing Challenge

Physical AI is not only a robotics story. It is a computing story at its core. As we outlined in our explainer on the emerging challenges of building hardware for Physical AI, bringing intelligence into real-world environments is redefining compute design, power efficiency, sensor integration, and verification workflows. Physical AI systems need hardware that can understand 3D environments, process multimodal data in real time, and operate reliably under strict latency and safety constraints.

To support this shift, NVIDIA organizes the development pipeline for physical AI around three distinct computers (each combining hardware, system software, and models):

1) one for training,
2) one for simulation and synthetic data generation,
3) and one for on-robot inference.

Together, these systems form a full stack for building, testing, and deploying robotic intelligence.

NVIDIA’s Three Computers for Physical AI

Each tier serves a specific purpose, provides dedicated hardware or software, and passes a defined output to the next stage of the pipeline.

Tier 1: Training Computer

Purpose: Enable large-scale training of robot foundation models for multimodal perception, language reasoning, and policy learning. Models must learn to understand objects, environments, and actions across diverse data modalities.

Architecture:

Hardware
- NVIDIA DGX systems provide the compute backbone for model training

Software and Models
- NVIDIA Cosmos open world foundation models
- NVIDIA Isaac GR00T humanoid robot foundation models

Developers can pre-train models from scratch on DGX or fine-tune Cosmos and GR00T for robot-specific tasks.

Output: A pre-trained or post-trained robotic foundation model ready for validation, testing, and refinement through simulation.

Tier 2: Simulation and Synthetic Data Computer

Purpose: Enable scalable simulation and synthetic data generation to support policy learning and robust validation of robot behavior when real-world data is limited, costly, or unsafe to collect.

Architecture:

Hardware
- NVIDIA RTX PRO Servers powering Omniverse-based simulation workloads

Software and Models
- NVIDIA Omniverse for physically based simulation
- NVIDIA Isaac Sim for robotics simulation
- NVIDIA Isaac Lab for reinforcement and imitation learning
- NVIDIA Cosmos for synthetic perception data

These systems allow developers to generate training data and validate robot policies across diverse environments and edge cases before real-world deployment.

Output: A validated robot policy tested across simulated scenarios and ready for real-time deployment on physical hardware.

Tier 3: Runtime Computer

Purpose: Enable real-time onboard inference for perception, reasoning, motion planning, and control. Robots must process sensor data, understand their environment, and act within milliseconds while operating under tight power and size constraints.

Architecture:

Hardware
- NVIDIA Jetson AGX Thor as the onboard AI compute system for real-time processing

Software and Models
- Runtime models for perception, control, and language
- Multimodal inference frameworks for robot decision-making

The system processes onboard sensor inputs, reasons about the physical environment, plans next actions, and executes motion on the robot.

Output: Real-time robotic behavior executed on physical hardware, with perception, planning, and control running directly on the onboard compute system.

You can find further publications of NVIDIA covering physical AI here.

Spotlights

ā€œFirgun Ventures announced a $70 million first close to launch what it calls the market’s first VC fund focused on early growth-stage quantum technology companies, backed by the Qatar Investment Authority. The fund, targeting $250 million, will invest globally in quantum computing, sensing, communications, and related applications in sectors such as healthcare, climate science, finance, and cybersecurity.

Founded by Kris Naudts and Zeynep Koruturk, the firm combines scientific, entrepreneurial, and financial expertise supported by an advisory council from leading institutions including Cambridge, Oxford, MIT, Google, the European Investment Bank, and the Wellcome Trust.ā€

If you want to learn what Zeynep and Kris look for in quantum startups and how they think about exit pathways, check out the interview on the Future of Computing blog!

🦾 The Energy Economics of Inference (Contrarian Ventures)

This deep dive explains that the real constraint in modern AI is increasingly energy per token rather than compute. As inference becomes sparse, memory-heavy and agentic, most energy is spent on memory access, data movement and idle time instead of FLOPs. This exposes a structural mismatch with GPUs, which were built for dense, regular workloads. The analysis points toward architectures designed around bandwidth and locality, including memory-centric chips, optical interconnects and neuromorphic systems that better match the event-driven nature of modern inference.

Headlines


Last week’s headlines featured new semiconductor developments from SK hynix, TSMC, and imec, advances in quantum, photonic, and neuromorphic hardware, major AI infrastructure moves in data centers and cloud, and the launch of Claude Opus 4.5.

āš›ļø Quantum

āš”ļø Photonic / Optical

🧠 Neuromorphic

šŸ’„ Data Centers

ā˜ļø Cloud

šŸ¤– AI

Readings


This reading list covers Europe’s diamond ambitions, open chiplet designs, quantum insights from diamond defects, nonlinear photonic architectures, neuromorphic systems with time sync and cooling, and the AI-driven shift in data center buildouts.

🦾 Semiconductors

Improving IC System Quality and Performance (SemiEngineering) (14 mins – Video)

āš›ļø Quantum

āš”ļø Photonic / Optical

🧠 Neuromorphic

šŸ’„ Data Centers

AWS has more than 900 data centers — report (Data Center Dynamics) (5 mins)

Funding News


After several very active weeks, this one was noticeably quieter and unusually concentrated in early-stage activity. It was essentially a week of photonics. Did we miss any rounds?

Amount

Name

Round

Category

€1M

Shinephi

Seed

Photonics

$4.5M

SPhotonix

Pre-Seed

Photonics

€10M

QphoX

Venture Round

Quantum

$6.5M

LightSpeed Photonics

Pre-Series A 

Photonics

$10M

Mastiska

Seed

Semiconductors

Bonus: Google’s TPU Push Is Becoming a Real Threat to NVIDIA

A few key points on Google’s aggressive push to commercialize TPUs and challenge NVIDIA’s dominance:

  • Google is now selling TPUs directly to multiple external firms, which represents a major shift in its commercialization strategy.

  • Claude 4.5 Opus and Gemini 3 (ā€œthe two best models in the worldā€) run mostly on Google TPUs and Amazon Trainium, not on NVIDIA chips.

  • Gemini 3 was trained entirely on TPUs, showing that TPU performance is competitive at the highest model quality levels.

  • Anthropic is deploying up to 1 million TPUs, with 400,000 purchased as hardware racks and 600,000 rented through GCP, positioning Google as a true merchant silicon vendor.

  • OpenAI has already lowered its NVIDIA TCO by about 30% simply by negotiating with the option to adopt TPUs, showing the pricing pressure Google can now apply.

  • Meta, SSI, xAI, and potentially OpenAI are now being targeted as external TPU customers, signalling a rapidly expanding TPU ecosystem beyond Google’s own workloads.

ā¤ļø Love these insights? Forward this newsletter to a friend or two. They can subscribe here.