MarinbadGPT

MarinbadGPT is a language model based on HuggingFaceTB's SmolLM-135M architecture, finely trained on a corpus of Marinbad games. The aim of this model is to generate games of Marinbad and play them against a human player.

Model Training

The training of MarinbadGPT was conducted on a high-performance computing infrastructure utilizing NVIDIA H100 GPUs, renowned for their power in the field of deep learning.

Training Configuration:

  • Infrastructure: 2x NVIDIA H100 (80GB HBM3)
  • Duration: 1 hour
  • Optimizer: AdamW
  • Learning Rate: 3e-4
  • Batch Size: Micro batch size of 128, with gradient accumulation steps of 8, resulting in an effective batch size of 1024
  • Warmup Steps: 100
Downloads last month
10
Safetensors
Model size
135M params
Tensor type
BF16
·
Inference API
Unable to determine this model's library. Check the docs .