Skip to content

Quick Start

Ahura Majdinasab edited this page Mar 6, 2026 · 3 revisions

Quick Start Guide

This guide gets PrismBench running locally with the current microservice stack.

Prerequisites

  • Python 3.12+
  • Docker with Compose
  • Git

Installation

1. Clone the repository

git clone https://github.com/CommissarSilver/PrismBench.git
cd PrismBench

2. Set up a Python environment (optional but recommended)

python -m venv .venv
source .venv/bin/activate  # Unix/macOS
# or
.\.venv\Scripts\activate   # Windows

3. Install dependencies

pip install -r requirements.txt

Configuration

1. Set up API keys

Create apis.key from the template:

cp apis.key.template apis.key

Use KEY=value format (no quotes), for example:

OPENAI_API_KEY=sk-your-openai-key-here
ANTHROPIC_API_KEY=your-anthropic-key-here
DEEPSEEK_API_KEY=your-key-here
TOGETHERAI_API_KEY=your-key-here
LOCAL_AI_BASE_URL=http://ollama:11434/

2. Select model settings

PrismBench agent configs are role-based files in configs/agents/*.yaml. Edit the agents you plan to use, for example configs/agents/challenge_designer.yaml:

role: challenge_designer
model_name: gpt-4o-mini
model_provider: openai
api_base: https://api.openai.com/v1/
model_params:
  temperature: 0.8
  max_tokens: 5120

Running PrismBench

Start services

From repository root:

docker compose -f docker/docker-compose.yaml up --build

This starts four microservices:

  • LLM Interface (http://localhost:8000)
  • Environment (http://localhost:8001)
  • Search (http://localhost:8002)
  • GUI (http://localhost:3000)

Run a complete evaluation

Use either:

  • the GUI at http://localhost:3000, or
  • the Search API (/initialize, /run, /tasks/{task_id}) at http://localhost:8002.

Next Steps

Troubleshooting

For common setup and runtime issues, see troubleshooting.


Related Pages

Next Steps

Core Concepts

Advanced Usage

Clone this wiki locally