Name and Version
llama-server full cuda docker v b7139
Looks like this might be related to #17381
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-server
Command line
Problem description & steps to reproduce
I am using latest docker version of llama-server (b7139)
And i am getting this error when trying to use openweebui with native tool calls:
parse: error parsing grammar: number of repetitions exceeds sane defaults, please reduce the number of repetitions
I didnt have this issue with previous versions (with the exact same setup), not sure exactly when it broke, as i update llama server package once a week.
First Bad Commit
#17381
Relevant log output