Skip to content

Chess Environment

A chess reinforcement learning environment for OpenEnv, powered by the moonfish chess engine.

Features

  • Full chess rules via python-chess library
  • Configurable opponent: moonfish engine, random moves, or self-play
  • Position evaluation: Uses moonfish's PSQT-based evaluation
  • Standard OpenEnv interface: reset(), step(), state

Quick Start

Using Docker

# Build the image
docker build -t chess-env:latest -f envs/chess_env/server/Dockerfile .

# Run the server
docker run -p 8000:8000 chess-env:latest

Using the Client

The client is async by default:

import asyncio
from chess_env import ChessEnv, ChessAction

async def main():
    async with ChessEnv(base_url="http://localhost:8000") as env:
        # Reset for a new game
        result = await env.reset()
        print(f"Starting position: {result.observation.fen}")
        print(f"Legal moves: {result.observation.legal_moves}")

        # Make a move
        result = await env.step(ChessAction(move="e2e4"))
        print(f"Reward: {result.reward}, Done: {result.done}")

        # Play until game ends
        while not result.done:
            move = result.observation.legal_moves[0]
            result = await env.step(ChessAction(move=move))

        print(f"Game result: {result.observation.result}")

asyncio.run(main())

For synchronous usage, use the .sync() wrapper:

from chess_env import ChessEnv, ChessAction

with ChessEnv(base_url="http://localhost:8000").sync() as env:
    result = env.reset()
    result = env.step(ChessAction(move="e2e4"))
    print(f"Reward: {result.reward}")

Observation Space

Field Type Description
fen str Board position in FEN notation
legal_moves List[str] Legal moves in UCI format
is_check bool Whether current player is in check
done bool Whether game has ended
reward float Reward for last action
result str Game result ("1-0", "0-1", "1/2-1/2")

Action Space

Field Type Description
move str UCI format move (e.g., "e2e4", "e7e8q")

Rewards

Outcome Reward
Win +1.0
Loss -1.0
Draw 0.0
Illegal move -0.1

Configuration

The environment supports these configuration options:

Parameter Default Description
opponent "moonfish" Opponent type: "moonfish", "random", or None
opponent_depth 2 Search depth for moonfish opponent
max_moves 500 Maximum half-moves before draw
agent_color None Agent color: "white", "black", or None (alternate each episode)
gamma 0.99 Discount factor for temporal credit assignment

Temporal Discounting

For RL training, the environment computes temporally discounted rewards at episode end. This helps with credit assignment in long games where only the final outcome is known.

When an episode ends, the terminal observation's metadata includes: - discounted_rewards: List of discounted rewards for each agent move - gamma: The discount factor used

The formula is r_t = γ^(T-1-t) × R_final where: - T = total agent moves - t = move index (0-indexed) - R_final = terminal reward (+1, -1, or 0)

Example for a 5-move win with γ=0.99:

Move 0: 0.99^4 × 1.0 = 0.961
Move 1: 0.99^3 × 1.0 = 0.970
Move 2: 0.99^2 × 1.0 = 0.980
Move 3: 0.99^1 × 1.0 = 0.990
Move 4: 0.99^0 × 1.0 = 1.000