Update README.md
Browse files
README.md
CHANGED
@@ -3,4 +3,15 @@ license: mit
|
|
3 |
---
|
4 |
# Distilled-RoBERTa
|
5 |
|
6 |
-
The DistilBERT model is a [RoBERTa](https://huggingface.co/deepset/roberta-base-squad2-distilled) model, which is trained on the SQuAD 2.0 training set, fine-tuned on the [NewsQA](https://huggingface.co/datasets/lucadiliello/newsqa) dataset.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
# Distilled-RoBERTa
|
5 |
|
6 |
+
The DistilBERT model is a [RoBERTa](https://huggingface.co/deepset/roberta-base-squad2-distilled) model, which is trained on the SQuAD 2.0 training set, fine-tuned on the [NewsQA](https://huggingface.co/datasets/lucadiliello/newsqa) dataset.
|
7 |
+
|
8 |
+
## Hyperparameters
|
9 |
+
```
|
10 |
+
batch_size = 16
|
11 |
+
n_epochs = 3
|
12 |
+
max_seq_len = 512
|
13 |
+
learning_rate = 2e-5
|
14 |
+
optimizer=Adam
|
15 |
+
weight_decay=0.01
|
16 |
+
embeds_dropout_prob = 0.1
|
17 |
+
```
|