Good evening!
Sorry, I've run into a problem.
I've been testing the new flags for llama.cpp from reasoning-budget, and unfortunately they are not compatible with llama-swap.
Here is the message that appears every time I add them.
Mi configuration is simple:
${server_cmd}
--model ${models_dir}/tesslate/omnicoder-9b-q8.gguf
--ctx-size 204800
--n-predict 16384
--top-p 0.95
--top-k 20
--min-p 0.0
--temp 0.6
--reasoning-budget 1000
--reasoning-budget-message "... thinking budget exceeded, let's answer now."
Good evening!
Sorry, I've run into a problem.
I've been testing the new flags for llama.cpp from reasoning-budget, and unfortunately they are not compatible with llama-swap.
Here is the message that appears every time I add them.
Mi configuration is simple: