Update README.md
Browse files
README.md
CHANGED
@@ -36,4 +36,5 @@ inference:
|
|
36 |
## training
|
37 |
|
38 |
- for inputs, the model was presented with the post title and the post selftext encoded as: `question: <post title> context: <post selftext>`. You may see better results if queries are posed in this fashion.
|
39 |
-
- The top two replies were aggregated and presented to the model as the output text.
|
|
|
|
36 |
## training
|
37 |
|
38 |
- for inputs, the model was presented with the post title and the post selftext encoded as: `question: <post title> context: <post selftext>`. You may see better results if queries are posed in this fashion.
|
39 |
+
- The top two replies were aggregated and presented to the model as the output text.
|
40 |
+
- Training for longer will be explored, but given that the dataset has 127k examples and the loss flatlines at 0.5 epochs this should be fairly viable.
|