Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit fde47a9

Browse files
authored
Merge branch 'dev' into test/harry-reorganize-api-folder
2 parents 650eb80 + b4164c6 commit fde47a9

File tree

2 files changed

+320
-110
lines changed

2 files changed

+320
-110
lines changed

engine/services/inference_service.cc

Lines changed: 15 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -66,9 +66,21 @@ cpp::result<void, InferResult> InferenceService::HandleChatCompletion(
6666
tokenizer->add_eos_token, tokenizer->add_generation_prompt);
6767
if (prompt_result.has_value()) {
6868
(*json_body)["prompt"] = prompt_result.value();
69-
Json::Value stops(Json::arrayValue);
70-
stops.append(tokenizer->eos_token);
71-
(*json_body)["stop"] = stops;
69+
if (json_body->isMember("stop")) {
70+
bool need_append = true;
71+
for (auto& s : (*json_body)["stop"]) {
72+
if (s.asString() == tokenizer->eos_token) {
73+
need_append = false;
74+
}
75+
}
76+
if (need_append) {
77+
(*json_body)["stop"].append(tokenizer->eos_token);
78+
}
79+
} else {
80+
Json::Value stops(Json::arrayValue);
81+
stops.append(tokenizer->eos_token);
82+
(*json_body)["stop"] = stops;
83+
}
7284
} else {
7385
CTL_ERR("Failed to render prompt: " + prompt_result.error());
7486
}

0 commit comments

Comments
 (0)