| # PP489 Optimization Variants Iter1 x TIGIT (YM_0988) |
| |
| ## Overview |
| |
| YM_0988 includes ABC001 against 2 TIGIT homologs. We explored several model hypothesis: (i) Does pre-training aid predicitivity and (ii) does the featurization of the input sequences matter. To test pretraining, we refer to `mata_descriptions` with the term **warm** to include pretraining, and **cold** to start from a randomly initialized seed. For featurization, we explored **label-encoded** sequences with a one-hot-encoder of amino acid identities, versus an **ESM-featurized** embedding to represent each sequence in the PPI. Optimization was performed on the human ortholog. |
| |
| ## Experimental details |
| |
| We studied the efficacy of generating binders with different model hyperparameters. This dataset includes 26726 unique scFvs and 2 unique target sequences. |
| |
| A more extensive methods section can be found in our publication [here](https://pmc.ncbi.nlm.nih.gov/articles/PMC12296056/). |
| |
| ## Misc dataset details |
| |
| We define the following binders: |
| |
| ### A-library (scFvs) |
| |
| There are several terms you can filter by: |
| |
| - `ABC001_WT_<i>`: These are WT replicates. |
| - `ABC001_label_encoded_cold`: Label encoded sequences with no pretraining |
| - `ABC001_label_encoded_warm`: Label encoded sequences with pretraining |
| - `ABC001_esm_cold`: ESM featurized sequences with no pretraining |
| - `ABC001_esm_warm`: ESM featurized sequences with pretraining |
|
|
| ### Alpha-library |
|
|
| - `TIGIT_22-137_POI-AGA2`: Human TIGIT |
| - `TIGIT_Mouse`: Mouse TIGIT |
|
|