Conversation
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
ℹ️ About Codex in GitHub
Your team has set up Codex to review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
Codex can also answer questions or update the PR. Try commenting
97b5dc5 to
4d4db1f
Compare
4d4db1f to
d6c76c1
Compare
| body.setdefault("trainingConfig", {})["outputModel"] = build_default_output_model(evaluator_id) | ||
|
|
||
| # Clean None fields to avoid noisy payloads | ||
| body = {k: v for k, v in body.items() if v is not None} |
There was a problem hiding this comment.
Bug: Debug Print Shows Incomplete Request Body
The debug print of the request body (lines 224-225) occurs before the body is fully constructed. After printing, the code adds evaluationDataset and outputModel fields (lines 226-231) and filters out None values (line 234). This means the printed body shows an incomplete/incorrect preview that doesn't match what will actually be sent to the API. The print statement should be moved after line 234 to show the final request body.
FIREWORKS_API_KEY=$PYROWORKS_DEV_API_KEY FIREWORKS_ACCOUNT_ID=pyroworks-dev FIREWORKS_API_BASE="https://dev.api.fireworks.ai" eval-protocol create rft --base-model accounts/fireworks/models/qwen3-0p6b --dataset-id svgbench-small --output-model svgbench-agent-small-bchen-2 --evaluator accounts/pyroworks-dev/evaluators/test-livesvgbench-test-svg-combined-evaluation1 --max-context-length 65536 --skip-validations --response-candidates-count 4 --batch-size 128000 --chunk-size 50 --epochs 4 --use-temporal-workflow --use-eval-v3 --max-output-tokens 32768 --max-inference-replica-count 4 --learning-rate 0.00003 --lora-rank 16
Note
Adds a
create rftCLI to launch Reinforcement Fine-tuning jobs on Fireworks with dataset upload/materialization, training/inference/W&B options, and persists evaluator traces for smoother workflows.createwith subcommandrftto create Reinforcement Fine-tuning jobs.--dataset-id,--dataset-jsonl,--dataset-builder,--dataset-display-name), training config (--base-model,--warm-start-from,--epochs,--batch-size,--learning-rate,--max-context-length,--lora-rank,--accelerator-count,--region,--output-model,--evaluation-dataset,--[no-]eval-auto-carveout), inference params (--temperature,--top-p,--top-k,--max-tokens,--n,--inference-extra-body), W&B (--wandb-*), and misc (--rft-job-id,--yes,--dry-run).eval_protocol/fireworks_rft.py):create rftusage.Written by Cursor Bugbot for commit d6c76c1. This will update automatically on new commits. Configure here.