|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
datasets: |
|
- Barishni-blinchik/uwbruh |
|
tags: |
|
- kawaii |
|
- cringe |
|
--- |
|
|
|
***Some cringe...* Oh well, hello!** |
|
I present GPT2 to you, but with a bit of kawaii. |
|
|
|
Chat template |
|
``` |
|
<|USER|> Hello <|ASSISTANT|> |
|
``` |
|
|
|
--- |
|
|
|
# Training Results |
|
|
|
The following metrics are from the latest training session of our model: |
|
|
|
## Overview |
|
|
|
- **Global Step:** 615 |
|
- **Training Loss:** 0.1303 |
|
|
|
## Detailed Metrics |
|
|
|
- **Training Runtime:** 413.1481 seconds |
|
- **Training Samples per Second:** 5.947 |
|
- **Training Steps per Second:** 1.489 |
|
- **Total Floating Point Operations (FLOs):** 641,994,522,624,000.0 |
|
- **Training Loss:** 0.13032278840134784 |
|
- **Epoch:** 3.0 |
|
|
|
--- |