🦾 Breaking the Memory Wall Pt. 2: A Closer Look at HBM (High Bandwidth Memory)

A Newsletter for Entrepreneurs, Investors, and Computing Geeks

Happy Monday! Here’s what’s inside this week’s newsletter:

  • Deep dive: The second part of our Memory Wall series takes a closer look at HBM (High Bandwidth Memory), explaining how HBM’s layered architecture, production process, and shifting market demand define its position in today’s high-performance memory landscape.

  • Spotlights: Q.ANT introduces its second-generation photonic Native Processing Unit for energy-efficient AI and HPC and d-Matrix partners with Alchip on a 3D DRAM-based inference accelerator that validates its in-memory compute approach on test silicon.

  • Headlines: Semiconductor updates from metrology to PCB and SRAM advances, expanding quantum–HPC integrations and defense strategies, progress in photonic pilots and acquisitions, neuromorphic sensing research, and large data center, networking, and AI build-outs and partnerships.

  • Readings: Semiconductor supply chain and packaging challenges, quantum error correction and simulation progress, renewed photonics investment, neuromorphic advances, and energy and sovereignty issues in data center and cloud infrastructure.

  • Funding news: A very active week with numerous Seed and Series A rounds across cloud, semiconductors, photonics, quantum, and data centers, while Lambda’s $1.5B Series E shows how a single large cloud round can dominate overall capital deployed.

  • Bonus: Washington’s latest semiconductor and AI moves, from slowing broad chip-import tariffs and approving controlled Nvidia Blackwell exports to the Gulf, to proposing SEMI Act tax credits and preparing new limits on state-level AI rules amid strong public support for AI safety.

Deep Dive: Breaking the Memory Wall Pt. 2 - A Closer Look at HBM (High Bandwidth Memory)

Building on last week’s analysis of FMC and Majestic Labs and their innovations in the memory stack, this issue turns to the memory technology at the core of all major AI accelerators: High Bandwidth Memory (HBM).

What is HBM

As outlined last week, each memory tier comes with its own trade-offs. SRAM provides high speed but offers very little density, while DRAM delivers high density but does not provide the bandwidth required for modern workloads.

The dominant choice today is on-package High Bandwidth Memory (HBM), which combines vertically stacked DRAM layers, very wide data interfaces, and integration on the same package as the compute die. Because it is positioned directly beside GPUs, TPUs, and custom AI processors, HBM functions as the closest and fastest external memory layer available to these devices. Taken together, these design features allow HBM to provide a balanced mix of capacity, bandwidth, and efficiency compared to other memory types.

Hence, HBM is a memory technology that delivers:

  • high bandwidth through thousands of I/O connections

  • high density through vertically stacked DRAM dies

  • low energy per bit moved through very short transfer distances

This combination enables the multi-terabyte bandwidth required by current accelerators.

HBM Manufacturing

HBM production spans two main process stages:

  • Front-end processing: converting standard DRAM wafers into HBM wafers by forming TSVs (vertical connections through the silicon) and adding bumps (solder contacts for die-to-die connection) on both sides of the die.

  • Back-end processing: assembling and packaging the stacked dies using flows that support efficient joint formation, manage heat dissipation, and limit warpage and bump stress.

Two cross-cutting topics shape the overall manufacturing flow:

  • Yield challenges: arise from the demands of 3D stacking, delivering power through TSVs, and managing heat in a structure that becomes more sensitive as layer counts increase.

  • Bonding tools: must provide sub-micron alignment and controlled pressure to attach each die layer accurately, since misalignment or uneven force can damage bumps and reduce overall yield.

HBM Buyers and Demand

Demand for HBM has expanded rapidly in step with the growth of AI accelerators. Even with the increasing presence of custom ASICs, Nvidia is expected to remain the largest user of HBM in 2027. Its roadmap drives this, with Rubin Ultra alone moving to 1TB of HBM per GPU. Broadcom follows, supported by rising volumes of TPU and MTIA devices. Additional demand comes from OpenAI and SoftBank projects, which contribute smaller but still noticeable increases. Amazon is also becoming one of the major purchasers and follows a strategy of sourcing HBM directly rather than relying solely on design partners, which helps reduce its overall cost.

More detailed, chip-level bit forecasts are available in the Accelerator Model by SemiAnalysis.

Spotlights

“Q.ANT today announced the availability of its next-generation Native Processing Unit: The Q.ANT NPU 2, with enhanced nonlinear processing capabilities to deliver orders-of-magnitude gains in energy efficiency and performance for AI and high-performance workloads. By performing nonlinear mathematics natively in light, the Q.ANT NPU 2 enables entirely new classes of AI and scientific applications including physical AI and advanced robotics, next-generation computer vision and industrial intelligence, physics-based simulation and scientific discovery. Q.ANT is offering its NPUs directly as a 19” Server Solution including x86 host processor and Linux operating system.”

Stay tuned for our upcoming interview with Q.ANT on our blog!

“d-Matrix, the pioneer in generative AI inference for data centers, and Alchip, the high-performance and AI infrastructure ASIC leader, are collaboratively developing the world’s first 3D DRAM-based datacenter inference accelerator that will eliminate the performance and cost bottlenecks constraining today’s AI infrastructure.

The joint initiative combines Alchip’s ASIC design expertise with d-Matrix’s digital in-memory compute platform architecture. The collaboration has already enabled a key technology, d-Matrix 3DIMC™, that is featured on d-Matrix Pavehawk™ test silicon and has been successfully validated in d-Matrix’s labs.”

Headlines


Last week’s headlines featured new semiconductor updates, fresh quantum–HPC collaborations, progress in photonic and neuromorphic hardware, major data center expansion plans, and new alliances across networking and AI.

🦾 Semiconductors

⚛️ Quantum

⚡️ Photonic / Optical

🧠 Neuromorphic

💥 Data Centers

📡 Networking

🤖 AI

Readings


This reading list covers semiconductor supply chain risks and packaging costs, quantum progress amid talent shortages, advances in photonic and neuromorphic chips, energy challenges in data centers, and Europe’s evolving stance on cloud sovereignty.

🦾 Semiconductors

Made in China 2025: Evaluating China’s Performance (U.S.-China Economic and Security Review Commission) (Page 5 to 8 - 14 mins)

Issues In Ramping Advanced Packaging (SemiEngineering) (11 mins – Video)

⚛️ Quantum

⚡️ Photonic / Optical

Mode-locking in a semiconductor photonic bandgap laser (Communications Physics) (53 mins)

🧠 Neuromorphic

💥 Data Centers

☁️ Cloud

🤖 AI

Funding News


Last week saw a very high number of rounds across the computing stack, with unusually dense activity at the Seed and Series A stages. Despite the large number of early-stage raises, overall capital was dominated by a single very large cloud round (Lambda’s $1.5B Series E).

Amount

Name

Round

Category

CHF150K

YQuantum

Grant

Quantum

$4.2M

SubImage

Seed

Cloud

$5.3M

Luminal

Seed

Semiconductors

$6M

UbiQD

Debt

Quantum

$10M

Hammerhead AI

Seed

Data Centers

$12M

PicoJool

Seed

Photonics

$13.6M

Datum

Seed

Cloud

€14.1M

AlixLabs

Series A

Semiconductors

€16M

NcodiN

Seed

Photonics

$25M

PowerLattice

Series A

Semiconductors

$34M

Apono

Series B

Cloud

$80M

Amperesand

Series A

Data Center

$140M

Celero Communications

Series A + B

Photonics

$1.5B

Lambda

Series E

Cloud

Bonus: U.S. Policy Signals for Chips and AI

This week brought several policy moves in Washington that shape both the chip industry and the broader AI landscape.

The administration is expected to slow down its plan for broad semiconductor-import tariffs, as officials assess whether the move could heighten tensions with China.

The U.S. also approved controlled exports of up to 35,000 NVIDIA Blackwell chips to the UAE and Saudi Arabia, subject to strict security and reporting requirements.

Statement on UAE and Saudi Chip Exports (U.S. Department of Commerce)

In Congress, the SEMI Investment Act was introduced to extend the 35% advanced manufacturing tax credit to U.S. producers of semiconductor materials, including substrates, thin films, and process chemicals.

At the same time, the White House is preparing an executive order aimed at limiting state-level AI regulation, even as a national poll shows that 80% of U.S. adults support maintaining federal rules focused on AI safety and data security.

❤️ Love these insights? Forward this newsletter to a friend or two. They can subscribe here.