Update README.md
Browse files
README.md
CHANGED
@@ -67,7 +67,7 @@ cos(a, b)
|
|
67 |
* Training was conducted in two stages:
|
68 |
1. In the first stage, the model was trained on Wikipedia texts (4 million texts) for three epochs.
|
69 |
<img src="https://github.com/FractalGPT/ModelEmbedderDistilation/blob/main/DistilSBERT/Train/1_st_en.JPG?raw=true" width=700 />
|
70 |
-
3. In the second stage, training was conducted on Wikipedia
|
71 |
<img src="https://github.com/FractalGPT/ModelEmbedderDistilation/blob/main/DistilSBERT/Train/2_st_en.JPG?raw=true" width=700 />
|
72 |
|
73 |
## Full Model Architecture
|
|
|
67 |
* Training was conducted in two stages:
|
68 |
1. In the first stage, the model was trained on Wikipedia texts (4 million texts) for three epochs.
|
69 |
<img src="https://github.com/FractalGPT/ModelEmbedderDistilation/blob/main/DistilSBERT/Train/1_st_en.JPG?raw=true" width=700 />
|
70 |
+
3. In the second stage, training was conducted on Wikipedia and dialog dataset for one epoch.
|
71 |
<img src="https://github.com/FractalGPT/ModelEmbedderDistilation/blob/main/DistilSBERT/Train/2_st_en.JPG?raw=true" width=700 />
|
72 |
|
73 |
## Full Model Architecture
|