mirror of
https://github.com/ggerganov/llama.cpp
synced 2026-04-26 05:21:53 +02:00
read arch from gguf.MODEL_ARCH
This commit is contained in:
parent
280dd2dcb7
commit
c56ec07a9a
@ -4909,8 +4909,7 @@ class Mamba2Model(TextModel):
|
||||
# Fail early for models which don't have a block expansion factor of 2
|
||||
# TODO: does this really matter?
|
||||
# skip the assertion for FalconH1 Model
|
||||
architectures = self.hparams.get("architectures")
|
||||
if architectures is None or architectures[0] != "FalconH1ForCausalLM":
|
||||
if self.model_arch != gguf.MODEL_ARCH.FALCON_H1:
|
||||
assert d_inner == 2 * d_model
|
||||
assert d_inner % head_dim == 0
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user