Update README.md
Browse files
README.md
CHANGED
@@ -43,15 +43,15 @@ inference:
|
|
43 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
44 |
should probably proofread and complete it, then remove this comment. -->
|
45 |
|
46 |
-
# distilgpt2-
|
47 |
|
48 |
-
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on
|
49 |
It achieves the following results on the evaluation set:
|
50 |
- Loss: 2.2461
|
51 |
|
52 |
## Model description
|
53 |
|
54 |
-
|
55 |
|
56 |
## Intended uses & limitations
|
57 |
|
@@ -63,6 +63,7 @@ More information needed
|
|
63 |
|
64 |
## Training procedure
|
65 |
|
|
|
66 |
### Training hyperparameters
|
67 |
|
68 |
The following hyperparameters were used during training:
|
|
|
43 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
44 |
should probably proofread and complete it, then remove this comment. -->
|
45 |
|
46 |
+
# distilgpt2-tiny-conversational
|
47 |
|
48 |
+
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on a parsed version of Wizard of Wikipedia. Persona alpha/beta framework designed for use with [ai-msgbot](https://github.com/pszemraj/ai-msgbot).
|
49 |
It achieves the following results on the evaluation set:
|
50 |
- Loss: 2.2461
|
51 |
|
52 |
## Model description
|
53 |
|
54 |
+
- it is a splendid model
|
55 |
|
56 |
## Intended uses & limitations
|
57 |
|
|
|
63 |
|
64 |
## Training procedure
|
65 |
|
66 |
+
- deepspeed
|
67 |
### Training hyperparameters
|
68 |
|
69 |
The following hyperparameters were used during training:
|