Model Card for Dayhoff

Dayhoff is an Atlas of both protein sequence data and generative language models — a centralized resource that brings together 3.34 billion protein sequences across 1.7 billion clusters of metagenomic and natural protein sequences (GigaRef), 46 million structure-derived synthetic sequences (BackboneRef), and 16 million multiple sequence alignments (OpenProteinSet). These models can natively predict zero-shot mutation effects on fitness, scaffold structural motifs by conditioning on evolutionary or structural context, and perform guided generation of novel proteins within specified families. Learning from metagenomic and structure-based synthetic data from the Dayhoff Atlas increased the cellular expression rates of generated proteins, highlighting the real-world value of expanding the scale, diversity, and novelty of protein sequence data.

The Dayhoff architecture is a hybrid of state-space Mamba layers and Transformer self-attention, interleaved with Mixture-of-Experts modules to maximize capacity while preserving efficiency. It natively handles long contexts, allowing both single sequences and unrolled MSAs to be modeled. Trained with an autoregressive objective in both N→C and C→N directions, Dayhoff supports order-agnostic infilling and scales to billions of parameters.

Model Details

Model Description

  • Developed by: Kevin K. Yang, Sarah Alamdari, Alex J. Lee, Kaeli Kaymak-Loveless, Samir Char, Garyk Brixi, Carles Domingo-Enrich, Chentong Wang, Suyue Lyu, Nicolo Fusi, Neil Tenenholtz, Ava P. Amini
  • Model type: Hybrid state-space-model transformer architecture with mixture-of-experts
  • License: MIT

Model Sources

Uses

Downstream Use

Dayhoff is intended for broad research use on protein language modeling. The model has been used and assessed on the following capabilities:

  1. Unconditional design of protein sequences
  2. Zero-shot mutation effect prediction on ProteinGym
  3. Designing scaffolds for structural motifs in sequence space on RFDiffusion and MotifBench
  4. Homolog conditioning with Dayhoff-3b-GR-HM and Dayhoff-3b-GR-HM-c

Bias, Risks, and Limitations

This model should not be used to generate anything that is not a protein sequence or a set of homologuous protein sequences. It is not meant for natural language or other biological sequences, such as DNA sequences. Not all sequences are guaranteed to be realistic. It remains difficult to generate high-quality sequences with no sequence homology to any natural sequence.

How to Get Started with the Model

Sample protein generation code:


import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, set_seed

set_seed(0)
torch.set_default_device("cuda")

model = AutoModelForCausalLM.from_pretrained('microsoft/Dayhoff-170m-UR90')
tokenizer = AutoTokenizer.from_pretrained('microsoft/Dayhoff-170m-UR90', trust_remote_code=True)


inputs = tokenizer(tokenizer.bos_token, return_tensors="pt", return_token_type_ids=False)

outputs = model.generate(inputs['input_ids'],max_length=50,do_sample=True)
sequence = tokenizer.batch_decode(outputs,skip_special_tokens=True)
print(sequence)

For detailed instructions on package usage, please refer to the README in model repo.

Evaluation

Results

See the preprint for the latest benchmark results and evaluations.

Model perplexity on held-out test sequences for Dayhoff models.

Model UniRef50 GigaRef Aligned homologs Unaligned homologs
170m-UR50 11.62 11.88
170m-UR90 11.52 11.85
170m-GR 13.67 9.36
170m-UR50-BRn 11.78 12.03
170m-UR50-BRq 11.67 11.91
170m-UR50-BRu 11.66 11.87
3b-UR90 8.95 9.64
3b-GR-HM 11.95 6.68 4.34 4.60
3b-GR-HM-c 10.11 9.21 3.57 3.56

Quality of generated sequences as measured by ESMFold pLDDT and scPerplexity. Dataset statistics are for 1024 randomly-sampled sequences. Model statistics are for 1024 generations at T=1 in the N-to-C direction.

Model or dataset pLDDT (mean ± s.d.) scPerplexity (mean ± s.d.)
Natural sequences
UniRef50 0.653 ± 0.196 9.45 ± 2.89
GigaRef-clusters 0.619 ± 0.199 9.69 ± 2.83
GigaRef-singletons 0.561 ± 0.201 10.07 ± 2.88
Generated sequences
170m-UR50 0.421 ± 0.132 11.97 ± 2.14
170m-UR90 0.407 ± 0.125 12.12 ± 2.14
170m-GR 0.422 ± 0.129 11.83 ± 2.12
170m-UR50-BRu 0.441 ± 0.157 11.71 ± 2.18
170m-UR50-BRq 0.434 ± 0.152 11.72 ± 2.24
170m-UR50-BRn 0.432 ± 0.131 11.77 ± 2.24
3b-UR90 0.454 ± 0.150 11.79 ± 2.38
3b-GR-HM 0.406 ± 0.126 11.50 ± 2.16
3b-GR-HM-c 0.423 ± 0.132 11.91 ± 2.18

ProteinGym zero-shot performance Spearman’s correlation coefficient on ProteinGym substitutions and indels.

Input Model Parameters Substitutions Indels
Single sequence 170m-UR50 170M 0.353 0.479
170m-UR90 170M 0.354 0.483
170m-GR 170M 0.199 0.292
170m-UR50-BRu 170M 0.341 0.476
170m-UR50-BRq 170M 0.356 0.477
170m-UR50-BRn 170M 0.341 0.478
3b-UR90 3B 0.394 0.497
3b-GR-HM 3B 0.328 0.423
3b-GR-HM-c 3B 0.417 0.466
Aligned homologs 3b-GR-HM-c 3B 0.368 NA
Unaligned homologs 3b-GR-HM-c 3B 0.372 0.401

RFDiffusion Benchmark Performance Motif scaffolding performance, problems solved, successes out of 100, and MotifBench score.

Problem 170m-UR50 170m-UR90 170m-GR 170m-UR50-BRn 170m-UR50-BRq 170m-UR50-BRu 3b-UR90 3b-GR-HM 3b-GR-HM-c EvoDiff-Seq
1PRW 62 72 81 95 91 90 94 81 79 82
1BCF 0 0 5 0 0 0 10 8 0 7
5TPN 0 0 0 0 0 0 0 0 0 0
5IUS 0 0 0 0 0 0 0 0 0 0
3IXT 12 17 12 14 18 12 18 11 14 20
5YUI 0 0 0 0 0 0 0 0 0 0
1QJG 0 0 0 0 0 0 0 0 0 0
1YCR 2 5 0 6 7 6 2 3 4 2
2KL8 0 1 0 1 0 1 1 1 1 1
7MRX_60 1 0 0 0 0 2 42 0 9 0
7MRX_85 0 0 0 0 0 0 19 1 1 0
7MRX_128 0 0 0 0 0 0 0 0 0 0
4JHW 0 0 0 0 0 0 0 0 0 0
4ZYP 0 0 0 0 0 1 0 0 0 0
5WN9 0 0 0 0 0 0 0 0 0 0
6VW1 1 1 1 0 0 1 0 0 0 0
5TRV_short 0 0 0 0 0 0 0 0 0 0
5TRV_med 0 0 0 0 0 0 0 0 0 0
5TRV_long 0 0 0 0 0 0 0 0 0 0
6E6R_short 2 2 1 3 3 2 14 7 8 6
6E6R_med 0 1 2 0 0 2 4 0 2 0
6E6R_long 0 1 0 0 0 1 3 0 1 0
6EXZ_short 0 0 0 0 0 0 0 0 0 0
6EXZ_med 0 0 0 0 0 0 0 0 0 0
6EXZ_long 0 0 0 0 0 0 0 0 0 0
Problems solved 6 8 6 5 4 10 10 7 9 6
Successes 80 100 102 119 119 118 207 112 119 118
Score 9.65 12.25 6.10 7.26 10.62 14.36 16.32 11.90 14.14 7.67

MotifBench Benchmark Performance Motif scaffolding performance, problems solved, successes out of 100, and MotifBench score.

Problem 170m-UR50 170m-UR90 170m-GR 170m-UR50-BRn 170m-UR50-BRq 170m-UR50-BRu 3b-UR90 3b-GR-HM 3b-GR-HM-c EvoDiff-Seq
01_1LDB 1 1 3 0 0 1 20 2 12 0
02_1ITU 4 33 4 1 1 4 37 57 48 0
03_2CGA 0 0 0 0 0 0 0 0 0 0
04_5WN9 0 0 0 0 0 0 0 0 0 0
05_5ZE9 0 1 21 0 0 0 16 40 9 0
06_6E6R 1 1 1 1 2 1 6 3 1 2
07_6E6R 0 0 0 2 0 0 2 0 0 0
08_7AD5 0 0 0 0 0 0 0 0 0 0
09_7CG5 0 0 0 0 0 0 0 0 0 0
10_7WRK 0 0 0 0 0 0 0 0 0 0
11_3TQB 4 11 3 4 3 7 40 8 26 0
12_4JHW 0 0 0 0 0 0 0 0 0 0
13_4JHW 0 0 0 0 0 0 0 0 0 0
14_5IUS 0 0 0 0 0 0 0 0 0 0
15_7A8S 0 0 0 0 0 0 0 0 0 0
16_7BNY 0 0 0 0 0 0 0 0 0 0
17_7DGW 0 0 0 0 0 0 0 0 0 0
18_7MQQ 0 0 0 0 0 0 0 0 0 0
19_7MQQ 0 0 0 0 0 0 0 0 0 0
20_7UWL 0 0 0 0 0 0 0 0 0 0
21_1B73 0 0 0 0 0 0 0 0 0 0
22_1BCF 0 0 3 0 0 0 20 9 0 19
23_1MPY 0 0 0 0 0 0 0 0 0 0
24_1QY3 0 0 0 0 0 0 0 0 0 0
35_2RKX 0 0 0 0 0 0 0 0 0 0
36_3B5V 0 0 0 0 0 0 0 0 0 0
37_4XOJ 0 0 0 0 0 0 0 0 0 0
28_5YUI 0 0 0 0 0 0 0 0 0 0
29_6CPA 0 0 0 0 0 0 0 0 0 0
30_7UWL 0 0 0 0 0 0 0 0 0 0
Problems 4 5 6 4 3 4 7 6 5 2
Successes 10 47 35 8 6 13 141 119 96 21
Score 2.33 2.92 4.33 2.75 2.17 2.75 8.36 4.96 4.48 1.58

Technical Specifications

Compute Infrastructure

  • 170M-parameter models: trained on 8 NVIDIA A100 or 8 NVIDIA H100 GPUs using Distributed Data Parallel.
  • 3B-parameter models: trained on 176 NVIDIA H100 GPUs using Fully Sharded Data Parallel in hybrid-shard mode.

Responsible AI Considerations

The intended use of this model is to generate high-quality, realistic, protein sequences or sets of homologous protein sequences. Generations can be designed from scratch or conditioned on partial sequences in both N→C and C→N directions.

The code and datasets released in this repository are provided for research and development use only. They are not intended for use in clinical decision-making or for any other clinical use, and the performance of these models for clinical use has not been established. You bear sole responsibility for any use of these models, data and software, including incorporation into any product intended for clinical use.

Citation

If you use the code, data, models, or results. please cite our preprint.

Downloads last month
12
Safetensors
Model size
170M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including microsoft/Dayhoff-170m-UR90