MarinbadGPT
MarinbadGPT is a language model based on HuggingFaceTB's SmolLM-135M architecture, finely trained on a corpus of Marinbad games. The aim of this model is to generate games of Marinbad and play them against a human player.
Model Training
The training of MarinbadGPT was conducted on a high-performance computing infrastructure utilizing NVIDIA H100 GPUs, renowned for their power in the field of deep learning.
Training Configuration:
- Infrastructure: 2x NVIDIA H100 (80GB HBM3)
- Duration: 1 hour
- Optimizer: AdamW
- Learning Rate: 3e-4
- Batch Size: Micro batch size of 128, with gradient accumulation steps of 8, resulting in an effective batch size of 1024
- Warmup Steps: 100
- Downloads last month
- 10