Skip to content

Commit c1b0516

Browse files
committed
try
1 parent ce04620 commit c1b0516

File tree

2 files changed

+6
-14
lines changed

2 files changed

+6
-14
lines changed

eval_protocol/pytest/default_single_turn_rollout_process.py

Lines changed: 0 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -5,20 +5,8 @@
55
from typing import List
66

77
from litellm import acompletion
8-
import litellm
98
from typing import Dict
109

11-
# Fix LiteLLM event loop binding issues by disabling the logging worker
12-
try:
13-
# Disable LiteLLM's async logging worker to prevent event loop binding issues
14-
import litellm.litellm_core_utils.logging_worker
15-
16-
# Monkey patch to disable the worker entirely
17-
original_start_worker = litellm.litellm_core_utils.logging_worker.start_worker
18-
litellm.litellm_core_utils.logging_worker.start_worker = lambda *args, **kwargs: None
19-
except Exception:
20-
pass # Best effort
21-
2210
from eval_protocol.dataset_logger import default_logger
2311
from eval_protocol.models import EvaluationRow, Message
2412
from openai.types import CompletionUsage
@@ -33,7 +21,6 @@ class SingleTurnRolloutProcessor(RolloutProcessor):
3321

3422
def __call__(self, rows: List[EvaluationRow], config: RolloutProcessorConfig) -> List[asyncio.Task[EvaluationRow]]:
3523
"""Generate single turn rollout tasks and return them for external handling."""
36-
3724
# Do not modify global LiteLLM cache. Disable caching per-request instead.
3825

3926
async def process_row(row: EvaluationRow) -> EvaluationRow:

eval_protocol/quickstart/llm_judge.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@
3838
# {
3939
# "model": "fireworks_ai/accounts/fireworks/models/qwen3-235b-a22b-instruct-2507",
4040
# },
41-
{"model": "gpt-4.1"},
41+
# {"model": "gpt-4.1"},
4242
{
4343
"max_tokens": 131000,
4444
"extra_body": {"reasoning_effort": "medium"},
@@ -49,6 +49,11 @@
4949
"extra_body": {"reasoning_effort": "low"},
5050
"model": "fireworks_ai/accounts/fireworks/models/gpt-oss-20b",
5151
},
52+
{
53+
"max_tokens": 131000,
54+
"extra_body": {"reasoning_effort": "low"},
55+
"model": "fireworks_ai/accounts/fireworks/models/gpt-oss-120b",
56+
},
5257
],
5358
rollout_processor=SingleTurnRolloutProcessor(),
5459
# preprocess_fn=split_multi_turn_rows,

0 commit comments

Comments
 (0)