wabu commited on
Commit
84dd429
·
verified ·
1 Parent(s): f8c84db

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -21
README.md CHANGED
@@ -18,9 +18,10 @@ AmpGPT2 is a language model capable of generating de novo antimicrobial peptides
18
 
19
  AmpGPT2 is a fine-tuned version of [nferruz/ProtGPT2](https://huggingface.co/nferruz/ProtGPT2) based on the GPT2 Transformer architecture.
20
 
21
- | Training Loss | Epoch | Validation Loss | Accuracy |
22
- |:-------------:|:-----:|:---------------:|:--------:|
23
- | 3.7948 | 50.0 | 3.9890 | 0.4213 |
 
24
 
25
  To validate the results the Antimicrobial Peptide Scanner vr.2 (https://www.dveltri.com/ascan/v2/ascan.html) was used, which is a deep learning tool specifically designed for AMP recognition.
26
 
@@ -62,24 +63,9 @@ The following hyperparameters were used during training:
62
  - lr_scheduler_type: linear
63
  - num_epochs: 50.0
64
 
65
- \begin{table}[h!]
66
- \centering
67
- \caption{AMP Yield Comparison between AmpGPT2 and ProtGPT2}
68
- \begin{tabular}{lccc}
69
- \toprule
70
- Model & Total Sequences & AMP Classified & AMP Percentage (AMP\%) \\
71
- \midrule
72
- AmpGPT2 & 10000 & 9541 & 95.41\% \\
73
- ProtGPT2 & 10000 & 5530 & 55.3\% \\
74
- \bottomrule
75
- \end{tabular}
76
- \label{tab:amp_yield}
77
- \end{table}
78
-
79
- | Model | Amp% | Length |
80
- |:-------:|:-----:|:-------:|
81
- |AmpGPT2|95.86|64.08 |
82
- |ProtGPT2| 51.85 | 222.59 |
83
 
84
  ### Framework versions
85
 
 
18
 
19
  AmpGPT2 is a fine-tuned version of [nferruz/ProtGPT2](https://huggingface.co/nferruz/ProtGPT2) based on the GPT2 Transformer architecture.
20
 
21
+ | Model | sequences generated | Amp% | average length |
22
+ |:------------:|:-----------:|:-----------:|:-----------:|
23
+ | AmpGPT2| 1000 | 95.86| 64.08 |
24
+ | ProtGPT2| 1000 | 51.85 | 222.59 |
25
 
26
  To validate the results the Antimicrobial Peptide Scanner vr.2 (https://www.dveltri.com/ascan/v2/ascan.html) was used, which is a deep learning tool specifically designed for AMP recognition.
27
 
 
63
  - lr_scheduler_type: linear
64
  - num_epochs: 50.0
65
 
66
+ | Training Loss | Epoch | Validation Loss | Accuracy |
67
+ |:-------------:|:-----:|:---------------:|:--------:|
68
+ | 3.7948 | 50.0 | 3.9890 | 0.4213 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
 
70
  ### Framework versions
71