From ae42182246dfb28da96798d4e54d2c96109ea4d9 Mon Sep 17 00:00:00 2001 From: Alexander Eichhorn Date: Tue, 7 Apr 2026 03:52:06 +0200 Subject: [PATCH] fix: detect Z-Image LoRAs with transformer.layers prefix (#8986) OneTrainer exports Z-Image LoRAs with 'transformer.layers.' key prefix instead of 'diffusion_model.layers.'. Add this prefix (and the PEFT-wrapped 'base_model.model.transformer.layers.' variant) to the Z-Image LoRA probe so these models are correctly identified and loaded. --- invokeai/backend/model_manager/configs/lora.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/invokeai/backend/model_manager/configs/lora.py b/invokeai/backend/model_manager/configs/lora.py index 1619c9d6f0..791ded2ed0 100644 --- a/invokeai/backend/model_manager/configs/lora.py +++ b/invokeai/backend/model_manager/configs/lora.py @@ -711,6 +711,8 @@ class LoRA_LyCORIS_ZImage_Config(LoRA_LyCORIS_Config_Base, Config_Base): state_dict, { "diffusion_model.layers.", # Z-Image S3-DiT layer pattern + "transformer.layers.", # OneTrainer/diffusers prefix variant + "base_model.model.transformer.layers.", # PEFT-wrapped variant }, ) @@ -747,6 +749,8 @@ class LoRA_LyCORIS_ZImage_Config(LoRA_LyCORIS_Config_Base, Config_Base): state_dict, { "diffusion_model.layers.", # Z-Image S3-DiT layer pattern + "transformer.layers.", # OneTrainer/diffusers prefix variant + "base_model.model.transformer.layers.", # PEFT-wrapped variant }, )