mirror of
https://github.com/ggerganov/llama.cpp
synced 2026-03-16 12:10:54 +01:00
context : fix SWA-related warning for multiple sequences (#14045)
This commit is contained in:
parent
d17a809ef0
commit
487a5e0401
@ -123,7 +123,7 @@ llama_context::llama_context(
|
||||
__func__, n_ctx_per_seq, hparams.n_ctx_train);
|
||||
}
|
||||
|
||||
if (!params.swa_full && cparams.n_seq_max > 1) {
|
||||
if (!params.swa_full && cparams.n_seq_max > 1 && hparams.is_swa_any()) {
|
||||
LLAMA_LOG_WARN("%s: requested n_seq_max (%u) > 1, but swa_full is not enabled -- performance may be degraded: %s\n",
|
||||
__func__, cparams.n_seq_max, "https://github.com/ggml-org/llama.cpp/pull/13845#issuecomment-2924800573");
|
||||
}
|
||||
|
||||
Loading…
Reference in New Issue
Block a user