| .. |
|
cmake
|
Merge mainline llama.cpp (#3)
|
2024-07-27 07:55:01 +02:00 |
|
base64.hpp
|
llava : expose as a shared library for downstream projects (#3613)
|
2023-11-07 00:36:23 +03:00 |
|
build-info.cpp.in
|
build : link against build info instead of compiling against it (#3879)
|
2023-11-02 08:50:16 +02:00 |
|
chat-parser-xml-toolcall.cpp
|
Add chat parser for MiroThinker (#1138)
|
2026-01-13 08:07:12 +02:00 |
|
chat-parser-xml-toolcall.h
|
Add chat parser for MiroThinker (#1138)
|
2026-01-13 08:07:12 +02:00 |
|
chat-parser.cpp
|
Add chat parser for MiroThinker (#1138)
|
2026-01-13 08:07:12 +02:00 |
|
chat-parser.h
|
Refactor chat and server file (#1062)
|
2025-12-15 08:27:20 +01:00 |
|
chat.cpp
|
llama : add token matching support to llama-grammar (#1220)
|
2026-02-03 07:57:17 +02:00 |
|
chat.h
|
Add chat parser for MiroThinker (#1138)
|
2026-01-13 08:07:12 +02:00 |
|
CMakeLists.txt
|
llama : add token matching support to llama-grammar (#1220)
|
2026-02-03 07:57:17 +02:00 |
|
common.cpp
|
Server: add string ban (#1185)
|
2026-02-05 08:12:34 +02:00 |
|
common.h
|
Server: add string ban (#1185)
|
2026-02-05 08:12:34 +02:00 |
|
console.cpp
|
check C++ code with -Wmissing-declarations (#3184)
|
2023-09-15 15:38:27 -04:00 |
|
console.h
|
gguf : new file format with flexible meta data (beta) (#2398)
|
2023-08-21 23:07:43 +03:00 |
|
json-partial.cpp
|
common: Generalized XML-style tool-call parsing with streaming support (#958)
|
2025-11-18 15:29:58 +01:00 |
|
json-partial.h
|
Move minja and nlohmann/json to vendor (#802)
|
2025-09-27 09:12:35 +02:00 |
|
json-schema-to-grammar.cpp
|
Update grammar (#1023)
|
2025-11-30 18:45:38 +01:00 |
|
json-schema-to-grammar.h
|
common: Generalized XML-style tool-call parsing with streaming support (#958)
|
2025-11-18 15:29:58 +01:00 |
|
llguidance.cpp
|
Tool calls support from mainline (#723)
|
2025-09-01 08:38:49 +03:00 |
|
log.cpp
|
Refactor chat and server file (#1062)
|
2025-12-15 08:27:20 +01:00 |
|
log.h
|
Server: refactor and rename functions (#1151)
|
2026-01-18 08:16:57 +02:00 |
|
ngram-cache.cpp
|
Fixed lookup compilation issues on Windows (#6273)
|
2024-03-24 14:21:17 +01:00 |
|
ngram-cache.h
|
Merge mainline llama.cpp (#3)
|
2024-07-27 07:55:01 +02:00 |
|
regex-partial.cpp
|
llama : add token matching support to llama-grammar (#1220)
|
2026-02-03 07:57:17 +02:00 |
|
regex-partial.h
|
Tool calls support from mainline (#723)
|
2025-09-01 08:38:49 +03:00 |
|
sampling.cpp
|
llama : add token matching support to llama-grammar (#1220)
|
2026-02-03 07:57:17 +02:00 |
|
sampling.h
|
llama : add token matching support to llama-grammar (#1220)
|
2026-02-03 07:57:17 +02:00 |
|
speculative.cpp
|
Server: refactor and rename functions (#1151)
|
2026-01-18 08:16:57 +02:00 |
|
speculative.h
|
Port universal assisted decoding to llama-server (#699)
|
2025-08-18 09:22:23 +03:00 |
|
train.cpp
|
Server: refactor and rename functions (#1151)
|
2026-01-18 08:16:57 +02:00 |
|
train.h
|
sync : ggml (backend v2) (#3912)
|
2023-11-13 14:16:23 +02:00 |