Skip to content

lazy-guy/chess-llama

Repository files navigation

Chess-Llama

We trained a tiny Llama-based decoder-only transformer model for chess play, consisting of 23M parameters. The model is trained on a 3 million high-quality chess games from the Lichess Elite Database, on a single Nvidia L4 GPU for 18 hours, using the Google Cloud’s Vertex AI platform.

View on Huggingface

Web Version

This model can be run within a browser, thanks to Huggingface transformers.js! You can try it here

Performance

It uses the UCI format for input and output. It has been trained with the token indicating result appended to the beginning of the games, hoping it would improve performance during actual chess play. The model achieves an estimated Elo rating of 1400, and easily outperforms Skill-level 0 Stockfish, but loses to Stockfish set to level higher than 1.

Analysis

About

Tiny Llama model trained to play chess

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published