metadata
license: apache-2.0
datasets:
- markusheimerl/socratic_dialogs
language:
- en
- 100 million trainable parameters
- GPT model definition from https://github.com/mistralai/mistral-src/blob/main/one_file_ref.py
- Trained for 3 hours on 4xA100 80GB.
- Repo at https://github.com/ratisbonrobotics/gpt