Skip to main content
Our foundation models span multiple scales of human biology — from molecular signals to whole-patient trajectories. All models are available on HuggingFace.

Featured: SMB-v1-Structure

SMB-v1-Structure

1.7B parameters · Our flagship biological world modelThe first multimodal foundation model for oncology built on Joint-Embedding Predictive Architecture (JEPA). Predicts patient trajectories, not tokens.

What Makes It Different

State Prediction

Predicts future patient states in latent space, not text tokens.

Causal Learning

Learns cause-and-effect from (Pre-State + Intervention) → Post-State.

Multimodal Fusion

Ingests genomics, imaging, EHR, and proteomics into unified embeddings.

Architecture

The Standard Model uses a Joint-Embedding Predictive Architecture (JEPA) — treating the patient as a dynamic “world” and treatments as interventions that change that world.
Standard Model Architecture

Model Families


SMB-EHR

Foundation models for electronic health records, trained to predict clinical events and understand patient disease trajectories.

smb-ehr-4b

4B parameters · EHR Foundation ModelReframes EHRs as timestamped chains of clinical events and predicts next events to improve temporal reasoning over disease trajectories.

SMB-Vision

Medical imaging foundation models trained on radiology and pathology data. These encoders power the vision capabilities of the Standard Model.

smb-vision

97M - 600M parameters · Medical Imaging Foundation ModelsVision encoders for radiology, pathology, and CT imaging tasks. Multiple sizes available from base (97M) to large (600M) variants.

SMB-Language

Biomedical language models for clinical text understanding and sentence similarity.

smb-mntp-llama-3.1-8b-v1

8B parameters · Sentence SimilarityFine-tuned Llama 3.1 for biomedical sentence similarity and text understanding.

Model Selection Guide

Use SMB-v1-Structure for predicting how patients will evolve over time, simulating treatment outcomes, and modeling disease dynamics.
Use smb-ehr-4b for next-event prediction, temporal reasoning over EHRs, and disease trajectory analysis from clinical records.
Start with smb-vision-base for general tasks. Use smb-vision-ct-base-0519 for CT-specific applications or smb-vision-v0-risk for risk stratification.
Use smb-mntp-llama-3.1-8b-v1 for semantic search, sentence similarity, and clinical text embeddings.

All Models

Complete catalog of available models:
ModelFamilyParametersTaskLink
SMB-v1-1.7B-StructureStandard Model1.7BWorld Model / Trajectory PredictionHuggingFace
smb-ehr-4bSMB-EHR4BEHR / Next Event PredictionHuggingFace
smb-vision-v0-riskSMB-Vision0.6BVision / Risk AssessmentHuggingFace
smb-vision-v0-mimSMB-Vision0.6BVision / Masked Image ModelingHuggingFace
smb-vision-largeSMB-Vision0.3BVision / General EncoderHuggingFace
smb-vision-baseSMB-Vision97.2MVision / General EncoderHuggingFace
smb-vision-ct-base-0519SMB-Vision97.2MVision / CT-SpecificHuggingFace
smb-vision-vjepa2-vitl-384-256SMB-Vision0.3BVision / V-JEPA2HuggingFace
smb-mntp-llama-3.1-8b-v1SMB-Language8BLanguage / Sentence SimilarityHuggingFace

Hardware Requirements

GPU memory requirements vary by model size. We recommend the following minimums:
Model SizeMinimum GPU MemoryRecommended
~100M (Base)4 GB8 GB
~300M (Large)8 GB16 GB
~600M (v0)12 GB24 GB
1.7B (SMB-v1-Structure)16 GB32 GB
4B (EHR)24 GB48 GB
8B (Language)32 GB80 GB

Next Steps