- Clone
ppiseqrepo. cd ppiseqpython -m pip install -U -e .
export CUDA_VISIBLE_DEVICES=0,1
# For full finetuning with BFloat16:
accelerate launch --mixed-precision bf16 train.py
# For ConvBERT training with frozen embeddings:
accelerate launch train.py
# Important Note:
# Set `ddp_find_unused_parameters` parameter to `true` when training ESM2 or ESM3,
# this parameter is in the `train` key in the training/config/config.yamlTable 1: Top 10 runs ranked by test set Spearman correlation (
| PLM | Setup | Spearman | Pearson | RMSE | Spearman | Pearson | RMSE |
|---|---|---|---|---|---|---|---|
| (Validation) | (Validation) | (Validation) | (Test) | (Test) | (Test) | ||
| ------------ | ----------------- | ---------------- | ---------------- | -------------- | ---------------- | ---------------- | -------------- |
| Prot-T5 | ConvBERT-PAD | 0.48 |
0.48 |
1.52 |
0.48 |
0.51 |
1.42 |
| Ankh2-Ext1 | Finetuning-HP | 0.48 |
0.49 |
1.50 |
0.47 |
0.48 |
1.45 |
| ESM2-650M | ConvBERT-HP | 0.44 |
0.43 |
1.86 |
0.47 |
0.48 |
1.74 |
| Ankh2-Ext2 | Finetuning-HP | 0.47 |
0.47 |
1.51 |
0.45 |
0.46 |
1.43 |
| ESM2-650M | Finetuning-HP | 0.45 |
0.44 |
1.75 |
0.44 |
0.45 |
1.68 |
| Ankh-Base | Finetuning-HP | 0.47 |
0.47 |
1.47 |
0.44 |
0.45 |
1.41 |
| Prot-T5 | ConvBERT-HP | 0.42 |
0.42 |
1.66 |
0.44 |
0.44 |
1.59 |
| Prot-T5 | Finetuning-PAD | 0.44 |
0.44 |
1.65 |
0.44 |
0.45 |
1.57 |
| Prot-T5 | Finetuning-HP | 0.47 |
0.46 |
1.51 |
0.44 |
0.45 |
1.47 |
| ESM2-3B | Finetuning-PAD | 0.45 |
0.45 |
1.49 |
0.44 |
0.46 |
1.43 |
Note: PAD: Pooled attention addition; HP: Hierarchical pooling