MarinbadGPT

MarinbadGPT is a language model based on HuggingFaceTB's SmolLM-135M architecture, finely trained on a corpus of Marinbad games. The aim of this model is to generate games of Marinbad and play them against a human player.

Model Training

The training of MarinbadGPT was conducted on a high-performance computing infrastructure utilizing NVIDIA H100 GPUs, renowned for their power in the field of deep learning.

Training Configuration:

  • Infrastructure: 2x NVIDIA H100 (80GB HBM3)
  • Duration: 1 hour
  • Optimizer: AdamW
  • Learning Rate: 3e-4
  • Batch Size: Micro batch size of 128, with gradient accumulation steps of 8, resulting in an effective batch size of 1024
  • Warmup Steps: 100
Downloads last month
24
Safetensors
Model size
135M params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.