Skip to content

Conversation

@loci-dev
Copy link

Mirrored from ggml-org/llama.cpp#17524

On server, we want to enable jinja by default to allow tool calling and default system prompts. More and more models require this, so I think it's finally time to make it enabled by default.

However, we don't want to enable this for other examples (like llama-cli or llama-run), because these examples cannot yet handle rolling back tokens. This can happen when the chat template wants to modify pass tokens, for example, one can delete the reasoning content from the formatted chat.

This PR also update the auto-generated docs via llama-gen-docs command

@loci-dev loci-dev force-pushed the main branch 27 times, most recently from 50d76f4 to cbd9848 Compare December 1, 2025 11:08
@loci-dev loci-dev force-pushed the main branch 30 times, most recently from 5dbcd6b to b63d258 Compare January 10, 2026 18:11
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants