Skip to main content
Standard Model Bio We are building the Standard Model for Biomedicine, a biological world model that moves beyond text prediction to model the causal dynamics of human disease.
New Release: SMB-v1-structure is now available — our flagship multimodal foundation model for oncology built on Joint-Embedding Predictive Architecture (JEPA).

The Problem with Current AI

For years, the approach to biomedical AI has been simple: take a massive general-purpose model, feed it medical text, and watch it pass licensing exams. But when tested on realistic patient cases requiring actual treatment decisions, GPT-4 achieves just 30.3% completeness. The disconnect lies in a single, flawed assumption: that language is a sufficient proxy for disease biology.
Human biology is not a linguistic problem — it is inherently multimodal. Today’s AI models have never “seen” an entire CT volume or “read” a whole-genome sequence; they have only processed text descriptions of them.

Our Approach: A Biological World Model

The Standard Model represents a paradigm shift:
Traditional LLMsStandard Model
Predicts tokensPredicts patient states
Static snapshotsDynamic trajectories
Text as proxyRaw biological signals
Pattern matchingCausal reasoning
Standard Model Architecture

From Description to Dynamics

A clinician does not simply ask “Is this cancer?” (classification). Rather, they ask:
“Given the patient’s current state S(t) and this specific intervention I, what will their state be in 6 months S(t+1)?”
Our Standard Model answers this question by modeling biological dynamics, not generating text.

Architecture Overview

The Standard Model is built on Joint-Embedding Predictive Architecture (JEPA) — the same paradigm powering advances in robotics and autonomous driving, now applied to human biology.
1

Modality Ingestion

Raw signals — genomics, proteomics, imaging, EHR data — pass through modality-specific encoders and fuse into a universal latent space.
2

State Prediction

Given the current patient state S(t) and an intervention A(t), the model predicts the future state S(t+1) in latent space.
3

Causal Learning

Training data is structured as (Pre-State + Intervention) → (Post-State), forcing the model to learn cause-and-effect relationships.
4

Hybrid Optimization

Supervised fine-tuning anchors the model to clinical outcomes while JEPA objectives learn the underlying dynamics.

Core Beliefs

Small differences in clinical efficacy relative to standard of care drive billions of dollars in revenue and change millions of lives. Every marginal improvement can define clinical, regulatory, or commercial success.
Biological signals exist at every scale — from molecular variations in genomic sequences to cellular structures in histopathology slides, to anatomical changes in CT volumes. Language alone cannot capture this complexity.
Clinical oncology is not a series of static snapshots; it is an evolving biological trajectory. We model how tumors progress and respond to therapy, not just what they look like at a single moment.
Patient data is inherently multimodal because patients are inherently multimodal. The best foundation models integrate data across modalities and disease areas.

Use Cases

Once the Standard Model learns a patient’s biological state, it can power diverse applications:

Treatment Simulation

Simulate how a tumor would evolve under Treatment A versus Treatment B.

Digital Twins

Create evolving digital representations for prognosis and intervention planning.

Clinical Trial Optimization

Forecast trials in silico and optimize inclusion/exclusion criteria.

Progression Prediction

Predict disease trajectories and time-to-event outcomes.

Response Prediction

Model probability of response to specific therapies.

Toxicity Assessment

Predict treatment-related adverse events before they occur.

Model Families

Next Steps

Stay Connected