mirror of
https://github.com/ggerganov/llama.cpp
synced 2026-03-13 18:50:48 +01:00
* Add DeepSeek V3.1 thinking mode support
- Added COMMON_CHAT_FORMAT_DEEPSEEK_V3_1 enum value
- Created common_chat_params_init_deepseek_v3_1() function (currently uses R1 implementation)
- Created common_chat_parse_deepseek_v3_1() function that handles V3.1 thinking format:
- Extracts reasoning content before '</think>' tag into reasoning_content
- Extracts regular content after '</think>' tag into content
- No opening '<think>' tag in V3.1 format
- Added detection logic for V3.1 templates based on pattern: 'message['prefix'] is defined and message['prefix'] and thinking'
- Added V3.1 case to parsing switch statement
This addresses the issue where V3.1 outputs reasoning content followed by '</think>' and then regular content without the opening '<think>' tag.
* Another attempt by V3.1 non-thinking
* Fix test, but it's not asserting anything.
* Ignore vim swap files in tests dir
* Update the test
* Try using try_find_literal instead of regex
* passing test
* Revert "Try using try_find_literal instead of regex"
This reverts commit
|
||
|---|---|---|
| .. | ||
| templates | ||
| .editorconfig | ||
| ggml-vocab-aquila.gguf | ||
| ggml-vocab-baichuan.gguf | ||
| ggml-vocab-bert-bge.gguf | ||
| ggml-vocab-bert-bge.gguf.inp | ||
| ggml-vocab-bert-bge.gguf.out | ||
| ggml-vocab-command-r.gguf | ||
| ggml-vocab-command-r.gguf.inp | ||
| ggml-vocab-command-r.gguf.out | ||
| ggml-vocab-deepseek-coder.gguf | ||
| ggml-vocab-deepseek-coder.gguf.inp | ||
| ggml-vocab-deepseek-coder.gguf.out | ||
| ggml-vocab-deepseek-llm.gguf | ||
| ggml-vocab-deepseek-llm.gguf.inp | ||
| ggml-vocab-deepseek-llm.gguf.out | ||
| ggml-vocab-falcon.gguf | ||
| ggml-vocab-falcon.gguf.inp | ||
| ggml-vocab-falcon.gguf.out | ||
| ggml-vocab-gpt-2.gguf | ||
| ggml-vocab-gpt-2.gguf.inp | ||
| ggml-vocab-gpt-2.gguf.out | ||
| ggml-vocab-gpt-neox.gguf | ||
| ggml-vocab-llama-bpe.gguf | ||
| ggml-vocab-llama-bpe.gguf.inp | ||
| ggml-vocab-llama-bpe.gguf.out | ||
| ggml-vocab-llama-spm.gguf | ||
| ggml-vocab-llama-spm.gguf.inp | ||
| ggml-vocab-llama-spm.gguf.out | ||
| ggml-vocab-mpt.gguf | ||
| ggml-vocab-mpt.gguf.inp | ||
| ggml-vocab-mpt.gguf.out | ||
| ggml-vocab-nomic-bert-moe.gguf | ||
| ggml-vocab-phi-3.gguf | ||
| ggml-vocab-phi-3.gguf.inp | ||
| ggml-vocab-phi-3.gguf.out | ||
| ggml-vocab-qwen2.gguf | ||
| ggml-vocab-qwen2.gguf.inp | ||
| ggml-vocab-qwen2.gguf.out | ||
| ggml-vocab-refact.gguf | ||
| ggml-vocab-refact.gguf.inp | ||
| ggml-vocab-refact.gguf.out | ||
| ggml-vocab-starcoder.gguf | ||
| ggml-vocab-starcoder.gguf.inp | ||
| ggml-vocab-starcoder.gguf.out | ||