--- license: apache-2.0 datasets: - markusheimerl/socratic_dialogs language: - en --- - 100 million trainable parameters - GPT model definition from https://github.com/mistralai/mistral-src/blob/main/one_file_ref.py - Trained for 3 hours on 4xA100 80GB. - Repo at https://github.com/ratisbonrobotics/gpt