From 212dbaf9a2b3b3cf6d5bbf7204feeafacb40c56f Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Thu, 14 Dec 2023 23:04:13 -0500 Subject: [PATCH] fix comment --- invokeai/backend/model_management/util.py | 2 +- invokeai/backend/model_manager/probe.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/invokeai/backend/model_management/util.py b/invokeai/backend/model_management/util.py index 441467a4c7..f4737d9f0b 100644 --- a/invokeai/backend/model_management/util.py +++ b/invokeai/backend/model_management/util.py @@ -59,7 +59,7 @@ def lora_token_vector_length(checkpoint: dict) -> int: lora_token_vector_length = _get_shape_1(key, tensor, checkpoint) elif key.startswith("lora_unet_") and ( "time_emb_proj.lora_down" in key - ): # recognizes format at https://civitai.com/models/224641 work + ): # recognizes format at https://civitai.com/models/224641 lora_token_vector_length = _get_shape_1(key, tensor, checkpoint) elif key.startswith("lora_te") and "_self_attn_" in key: tmp_length = _get_shape_1(key, tensor, checkpoint) diff --git a/invokeai/backend/model_manager/probe.py b/invokeai/backend/model_manager/probe.py index 25120e2e33..64eefb774e 100644 --- a/invokeai/backend/model_manager/probe.py +++ b/invokeai/backend/model_manager/probe.py @@ -401,7 +401,7 @@ class LoRACheckpointProbe(CheckpointProbeBase): elif token_vector_length == 1024: return BaseModelType.StableDiffusion2 elif token_vector_length == 1280: - return BaseModelType.StableDiffusionXL # recognizes format at https://civitai.com/models/224641 work + return BaseModelType.StableDiffusionXL # recognizes format at https://civitai.com/models/224641 elif token_vector_length == 2048: return BaseModelType.StableDiffusionXL else: