-
Notifications
You must be signed in to change notification settings - Fork 10.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Misc. bug: Missing <think> tag in response (DeepSeek R1) #11861
Comments
I observed the same problem when I was playing with non-thinking models and making them think within |
I had the same issue, but once I upgraded to a release greater than b4706, the issue went away. Looks like PR11607 resolved the problem. I get both of my think tags (<think> </think>) Here is how I am calling it (using shared library via llama-cpp-python:
|
Name and Version
I don't know whether it's a bug or not.
The latest Jinja chat template for the DeepSeek r1 model adds a
<think>\n
postfix to force the model into thinking.However, this makes all the responses losing the heading
<think>
tag like this:I suggest manually adding the
<think>
prefix in response whenadd_generation_prompt = true
.Operating systems
Linux
Which llama.cpp modules do you know to be affected?
libllama (core library)
Command line
Problem description & steps to reproduce
llama-server
DeepSeek R1
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: