Niksa Praljak
commited on
Commit
·
1f2e18d
1
Parent(s):
07026ee
Load pretrained weights and cleanup config for stage 1
Browse files- stage1_config.json +3 -3
- weights/LLMs/README.md +7 -0
stage1_config.json
CHANGED
@@ -28,14 +28,14 @@
|
|
28 |
"base_lr": 0.0016,
|
29 |
"global_batch_size": 80,
|
30 |
"lr": 0.0005,
|
31 |
-
"seq_model_path": "/project/ranganathanr/niksapraljak/
|
32 |
"pretrained_seq": true,
|
33 |
"trainable_seq": true,
|
34 |
"rep_layer": 33,
|
35 |
"protein_encoder_embedding": 1280,
|
36 |
"protein_encoder_lr": 0.0005,
|
37 |
"pLM_n_layers_to_finetune": 1,
|
38 |
-
"text_model_path": "/project/ranganathanr/niksapraljak/
|
39 |
"pretrained_text": true,
|
40 |
"trainable_text": true,
|
41 |
"text_encoder_embedding": 768,
|
@@ -47,4 +47,4 @@
|
|
47 |
"head_lr": 0.0005,
|
48 |
"inference_data_path": "/project/ranganathanr/niksapraljak/BioM3_PDZ/data/test_prompts_PDZ_swissprot_pfam_dataset.csv",
|
49 |
"inference_output_path": "/project/ranganathanr/niksapraljak/BioM3_PDZ/outputs/Stage1_test_prompts_PDZ.pt"
|
50 |
-
}
|
|
|
28 |
"base_lr": 0.0016,
|
29 |
"global_batch_size": 80,
|
30 |
"lr": 0.0005,
|
31 |
+
"seq_model_path": "/project/ranganathanr/niksapraljak/HF_repo/HF_BioM3_project/BioM3/weights/LLMs/esm2_t33_650M_UR50D.pt",
|
32 |
"pretrained_seq": true,
|
33 |
"trainable_seq": true,
|
34 |
"rep_layer": 33,
|
35 |
"protein_encoder_embedding": 1280,
|
36 |
"protein_encoder_lr": 0.0005,
|
37 |
"pLM_n_layers_to_finetune": 1,
|
38 |
+
"text_model_path": "/project/ranganathanr/niksapraljak/HF_repo/HF_BioM3_project/BioM3/weights/LLMs/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext",
|
39 |
"pretrained_text": true,
|
40 |
"trainable_text": true,
|
41 |
"text_encoder_embedding": 768,
|
|
|
47 |
"head_lr": 0.0005,
|
48 |
"inference_data_path": "/project/ranganathanr/niksapraljak/BioM3_PDZ/data/test_prompts_PDZ_swissprot_pfam_dataset.csv",
|
49 |
"inference_output_path": "/project/ranganathanr/niksapraljak/BioM3_PDZ/outputs/Stage1_test_prompts_PDZ.pt"
|
50 |
+
}
|
weights/LLMs/README.md
CHANGED
@@ -12,9 +12,16 @@ This folder contains the pre-trained weights for the **ESM2** and **PubMedBERT**
|
|
12 |
To download the ESM2 (650M parameter) model weights:
|
13 |
```bash
|
14 |
wget https://dl.fbaipublicfiles.com/fair-esm/models/esm2_t33_650M_UR50D.pt
|
|
|
15 |
```
|
16 |
|
17 |
### PubMedBERT Model
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
To download the PubMedBERT model weights:
|
19 |
```bash
|
20 |
git clone https://huggingface.co/microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext
|
|
|
12 |
To download the ESM2 (650M parameter) model weights:
|
13 |
```bash
|
14 |
wget https://dl.fbaipublicfiles.com/fair-esm/models/esm2_t33_650M_UR50D.pt
|
15 |
+
wget https://dl.fbaipublicfiles.com/fair-esm/regression/esm2_t33_650M_UR50D-contact-regression.pt
|
16 |
```
|
17 |
|
18 |
### PubMedBERT Model
|
19 |
+
|
20 |
+
Make sure large-file storage capabilities are installed in your environment before cloning HuggingFace model card.
|
21 |
+
```bash
|
22 |
+
git lfs install
|
23 |
+
```
|
24 |
+
|
25 |
To download the PubMedBERT model weights:
|
26 |
```bash
|
27 |
git clone https://huggingface.co/microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext
|