English
gpt_100m / README.md
markusheimerl's picture
Update README.md
5bc9d0c verified
|
raw
history blame contribute delete
No virus
304 Bytes
---
license: apache-2.0
datasets:
- markusheimerl/socratic_dialogs
language:
- en
---
- 100 million trainable parameters
- GPT model definition from https://github.com/mistralai/mistral-src/blob/main/one_file_ref.py
- Trained for 3 hours on 4xA100 80GB.
- Repo at https://github.com/ratisbonrobotics/gpt