11# Evaluation Framework Configuration
22# This configuration is shared across all evaluation runner scripts
3- # Supports OpenAI, Groq, OpenRouter, and LiteLLM providers
3+ # Example configuration for Cerebras models
44
55# API endpoint for the evaluation server
66api_endpoint : " http://localhost:8080"
@@ -9,28 +9,27 @@ api_endpoint: "http://localhost:8080"
99# These models are sent to the agent for processing requests
1010
1111main_model :
12- provider : " openai "
13- model_name : " gpt-5-mini "
14- api_key : " ${OPENAI_API_KEY }"
12+ provider : " cerebras "
13+ model_name : " zai-glm-4.6 "
14+ api_key : " ${CEREBRAS_API_KEY }"
1515
1616mini_model :
17- provider : " openai "
18- model_name : " gpt-5-nano "
19- api_key : " ${OPENAI_API_KEY }"
17+ provider : " cerebras "
18+ model_name : " zai-glm-4.6 "
19+ api_key : " ${CEREBRAS_API_KEY }"
2020
2121nano_model :
22- provider : " openai "
23- model_name : " gpt-5-nano "
24- api_key : " ${OPENAI_API_KEY }"
22+ provider : " cerebras "
23+ model_name : " zai-glm-4.6 "
24+ api_key : " ${CEREBRAS_API_KEY }"
2525
2626# Model configuration for judging evaluation responses
2727# This model is used locally to assess the quality of agent responses
2828
2929judge_model :
30- provider : " openai"
31- model_name : " gpt-5"
32- api_key : " ${OPENAI_API_KEY}"
33-
30+ provider : " cerebras"
31+ model_name : " zai-glm-4.6"
32+ api_key : " ${CEREBRAS_API_KEY}"
3433
3534# Execution settings
3635
0 commit comments