mirror of
https://github.com/ggerganov/llama.cpp
synced 2026-03-02 13:19:27 +01:00
* server: introduce self-speculative decoding * server: moved self-call into speculative.cpp * can_speculate() includes self-speculation Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * server: can_speculate() tests self-spec * server: replace can_speculate() with slot.can_speculate() Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * common: use %zu format specifier for size_t in logging Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * server: can_speculate() requires a task instance * common: ngram map, config self-speculative decoding * common: add enum common_speculative_type * common: add vector of speculative states * common: add option --spec-draftless * server: cleanup (remove slot.batch_spec, rename) * common: moved self-spec impl to ngram-map * common: cleanup (use common_speculative_state_draft) * spec : refactor * cont : naming * spec: remove --spec-config * doc: (draftless) speculative decoding * common: print performance in spec decoding * minor : cleanup * common : better names * minor : cleanup + fix build * minor: comments * CODEOWNERS: add common/ngram-map.* (#18471) * common : rename speculative.draftless_type -> speculative.type * ngram-map : fix uninitialized values * ngram-map : take into account the input can become shorter * ngram-map : revert len check for now * arg : change `--spec-draftless` -> `--spec-type` * spec : add common_speculative_state::accept() * spec : refactor + add common_speculative_begin() * spec : fix begin() call with mtmd * spec : additional refactor + remove common_speculative_params --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> |
||
|---|---|---|
| .. | ||
| batched | ||
| batched.swift | ||
| convert-llama2c-to-ggml | ||
| debug | ||
| deprecation-warning | ||
| diffusion | ||
| embedding | ||
| eval-callback | ||
| gen-docs | ||
| gguf | ||
| gguf-hash | ||
| idle | ||
| llama.android | ||
| llama.swiftui | ||
| lookahead | ||
| lookup | ||
| model-conversion | ||
| parallel | ||
| passkey | ||
| retrieval | ||
| save-load-state | ||
| simple | ||
| simple-chat | ||
| simple-cmake-pkg | ||
| speculative | ||
| speculative-simple | ||
| sycl | ||
| training | ||
| CMakeLists.txt | ||
| convert_legacy_llama.py | ||
| json_schema_pydantic_example.py | ||
| json_schema_to_grammar.py | ||
| llama.vim | ||
| pydantic_models_to_grammar_examples.py | ||
| pydantic_models_to_grammar.py | ||
| reason-act.sh | ||
| regex_to_grammar.py | ||
| server_embd.py | ||
| server-llama2-13B.sh | ||
| ts-type-to-grammar.sh | ||