This guide gets PrismBench running locally with the current microservice stack.
- Python 3.12+
- Docker with Compose
- Git
git clone https://github.com/CommissarSilver/PrismBench.git
cd PrismBenchpython -m venv .venv
source .venv/bin/activate # Unix/macOS
# or
.\.venv\Scripts\activate # Windowspip install -r requirements.txtCreate apis.key from the template:
cp apis.key.template apis.keyUse KEY=value format (no quotes), for example:
OPENAI_API_KEY=sk-your-openai-key-here
ANTHROPIC_API_KEY=your-anthropic-key-here
DEEPSEEK_API_KEY=your-key-here
TOGETHERAI_API_KEY=your-key-here
LOCAL_AI_BASE_URL=http://ollama:11434/PrismBench agent configs are role-based files in configs/agents/*.yaml.
Edit the agents you plan to use, for example configs/agents/challenge_designer.yaml:
role: challenge_designer
model_name: gpt-4o-mini
model_provider: openai
api_base: https://api.openai.com/v1/
model_params:
temperature: 0.8
max_tokens: 5120From repository root:
docker compose -f docker/docker-compose.yaml up --buildThis starts four microservices:
- LLM Interface (
http://localhost:8000) - Environment (
http://localhost:8001) - Search (
http://localhost:8002) - GUI (
http://localhost:3000)
Use either:
- the GUI at
http://localhost:3000, or - the Search API (
/initialize,/run,/tasks/{task_id}) athttp://localhost:8002.
- Configuration Overview - Detailed configuration options
- Architecture Overview - Service and data flow details
- MCTS Algorithm - Search process details
- Results Analysis - Interpreting outputs
For common setup and runtime issues, see troubleshooting.
- Configuration Overview - Detailed configuration options
- Architecture Overview - Understanding system components
- Agent System - How agents work together
- MCTS Algorithm - Understanding the core algorithm
- Environment System - Evaluation environments
- Results Analysis - Interpreting your results
- Extending PrismBench - Customizing the framework
- Custom Agents - Creating specialized agents
- troubleshooting - Common issues and solutions