🧬 🦾 The Computational Demands of Protein Science

A Newsletter for Entrepreneurs, Investors, and Computing Geeks

Happy Monday! Here’s what’s inside this week’s newsletter:

  • Deep dive: Protein science as a new compute workload, why structure prediction is so demanding, and how GPUs and specialized hardware are being adapted.

  • Spotlights: Apple’s new A19 chips with upgraded cooling and connectivity, and first insights from Mira Murati’s Thinking Machines Lab on making large-scale AI inference reproducible

  • Headlines: New semiconductor launches, IPO moves and physics breakthroughs in quantum, expansions in photonics, nuclear debates in data centers, cloud partnerships, shifting alliances in AI, and much more.

  • Readings: Chip bottlenecks and DRAM evolution, advances in quantum key distribution, photonic accelerators for AI, neuromorphic efficiency gains, and global data center, cloud trends, and much more.

  • Funding news: Multi-billion rounds from PsiQuantum and Mistral, alongside early to mid-stage activity in semiconductors and photonics with most deals coming in under $60M.

  • Bonus: A set of new neuromorphic market forecasts, covering chips, hardware, and sensors, all pointing toward steady growth and rising relevance across AI and next-gen computing.

Deep Dive: The Computational Demands of Protein Science

An episode from the NVIDIA AI Podcast last week (From AlphaFold to MMseqs2-GPU: How AI is Accelerating Protein Science) inspired me to dig deeper into how biology is emerging as a new class of workloads that shape hardware requirements.

Predicting protein structures has become one of the most important computational problems of our time. It enables drug discovery and enzyme engineering. Yet the workloads that power this science are not only biologically complex but also some of the most compute-intensive tasks ever attempted outside of physics.

Why Protein Structure Prediction is Computationally Demanding

At the core of protein inference are usually two demanding stages: 1) generating multiple sequence alignments (MSAs), which search through massive protein databases to find and align evolutionarily related sequences, and 2) running deep learning models such as AlphaFold and OpenFold, which use this information to predict the final 3D structure. The first stage captures evolutionary context from billions of sequences, while the second relies on deep neural networks with tens to hundreds of millions of parameters.

  • MSA bottleneck: Generating alignments for a given protein involves searching and matching across protein databases that contain hundreds of millions to billions of sequences. Even on modern GPUs, this step can dominate runtime (up to 80%) if not carefully optimized.

  • Inference bottleneck: Once alignments are ready, the models themselves require enormous matrix multiplications, attention layers, and symmetry-aware geometry operations. These workloads are GPU-bound, resembling large language models in structure but with additional symmetry constraints from physics.

In the real world, you rarely care about just one protein. Drug discovery means running predictions for thousands of candidate proteins, and other fields like enzyme engineering may require testing entire sets of designed or natural variants.

This multiplies both bottlenecks by orders of magnitude, which is why even GPU clusters start to strain.

What Protein Workloads Demand from Hardware

Protein science workloads may look similar to AI or HPC, but they stress the hardware stack in distinct ways. Making them tractable requires processors tuned for these demands:

  • High-Bandwidth Memory (HBM): Protein inference involves processing large alignment matrices and ensembles of candidate structures. Storing these in HBM provides bandwidth on the order of hundreds of GBs up to several TBs per second, which helps avoid I/O bottlenecks that would stall standard GPUs.

  • Massive Parallelism: GPUs are inherently parallel, but biology workloads push this further. Tensor operations in protein models and redesigned sequence alignment algorithms demand extreme thread-level parallelism to reach usable performance.

  • Multi-Instance Capability: A single GPU can be partitioned into multiple logical devices, each with dedicated compute and memory resources. This makes it possible to run several predictions in parallel on one chip, which is essential when screening thousands of proteins or processing large datasets efficiently.

Beyond raw hardware, software optimizations are required to unlock these features. Modern GPUs include specialized instructions, such as Tensor Core operations, which libraries expose to biology models. This hardware–software co-design is essential for symmetry-aware computations, delivering speedups while preserving accuracy.

How NVIDIA’s Latest GPU Targets Biology Workloads

NVIDIA just launched the RTX PRO 6000 Blackwell Server Edition GPU, showing up to 138Ɨ faster protein inference with MMseqs2-GPU (GPU-accelerated sequence alignment software) and OpenFold2 (open-source structure prediction model). With 96 GB of high-bandwidth memory and new software optimizations, it enables proteome-scale folding on a single server instead of a supercomputing cluster.

Further interesting sources:

The NVIDIA RTX PRO 6000 Blackwell Server Edition GPU sets a new benchmark for protein structure inference. Source: NVIDIA.

Spotlights

Likely built on TSMC’s N3P node, the new chips feature upgraded CPU/GPU cores with integrated Neural Accelerators for improved AI and graphics performance. The Pro models introduce a vapor-chamber cooling system, promising 20Ɨ better heat dissipation than previous designs. Apple is also deepening vertical integration with its own N1 networking chip and an updated C1X modem for more efficient connectivity.

šŸ¤– Defeating Nondeterminism in LLM Inference (Thinking Machines)

Mira Murati’s $2B-backed research lab Thinking Machines Lab, staffed with former OpenAI researchers, has published a first look into its work on reproducible AI. The blog post explores why large language models return different answers even when using deterministic settings like temperature zero, challenges the common ā€œconcurrency + floating pointā€ explanation, and outlines how inference engines might be redesigned to achieve truly reproducible results.

Headlines


Last week’s headlines brought new chip launches and policy reviews in semiconductors, IPO moves and physics breakthroughs in quantum, updates in photonics, nuclear debates and big build-outs in data centers, major cloud deals, and shifting alliances in AI.

On another note: The European Commission asked stakeholders to evaluate and review the Chips Act.

āš›ļø Quantum

āš”ļø Photonic / Optical

šŸ’„ Data Centers

ā˜ļø Cloud

šŸ¤– AI

Overview of recent CoreWeave developments. Source: Tech in Asia (article on CoreWeave’s new AI fund).

Readings


This week’s reading list covers chip bottlenecks and DRAM evolution, advances in quantum key distribution, photonic accelerators for AI inference, neuromorphic efficiency gains, and shifting dynamics in global data center and cloud markets.

🦾 Semiconductors

What Do LLMs Want From Hardware? (SemiEngineering) (16 mins)

āš›ļø Quantum

āš”ļø Photonic / Optical

🧠 Neuromorphic

šŸ’„ Data Centers

Top 10: Data Centre Companies in Europe (Data Centre Magazine) (6 mins)

ā˜ļø Cloud

Overview of Huawei chip ecosystem. Source: SemiAnalysis.

Funding News


Last week’s rounds underscored the scale divide in computing. AI and quantum dominated the top end with PsiQuantum ($2B) and Mistral ($1.7B). At the same time, semiconductors and photonics saw a string of early to mid-stage rounds below $60M. The pattern reflects capital concentrating in a few platform bets, while component-level innovation progresses in smaller steps.

Amount

Name

Round

Category

Undisclosed

Asperitas

Venture Round 

Data Centers

€6.5M

Agate Sensors

Seed + Grant

Photonics

$8M

Astrus

Seed

Semiconductors

$14M

Opticore

Seed

Photonics

$20.5M

Clockwork

Venture Round 

Semiconductors

$51M

proteanTecs

Series D

Semiconductors

$58M

Scintil Photonics

Series B

Photonics

$64.6M

DataCrunch

Series A

Cloud

$200M

Perplexity

Venture Round

AI

$230M

QuEra Computing

Series B

Quantum

€600M

EcoDataCenter

Debt

Data Centers

$1.7B

Mistral AI

Series C

AI

$2B

PsiQuantum

Series E

Quantum

Bonus: Neuromorphic by the Numbers

A number of fresh market reports on neuromorphic computing were published last week. Different scopes, geographies, and methodologies, but one common theme: steady growth across chips, hardware, and sensors. Treat the figures as directional, since each report uses its own assumptions.

ā¤ļø Love these insights? Forward this newsletter to a friend or two. They can subscribe here.