Skip to content

GemmaLoader Transformers does not recognize this architecture. #114

@Alexander-RPR

Description

@Alexander-RPR

GemmaLoader
The checkpoint you are trying to load has model type gemma2 but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

ComfyUI Error Report

Error Details

  • Node ID: 8
  • Node Type: GemmaLoader
  • Exception Type: ValueError
  • Exception Message: The checkpoint you are trying to load has model type gemma2 but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

Stack Trace

  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)

  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
    text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
    config, kwargs = AutoConfig.from_pretrained(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
    raise ValueError(

System Information

  • ComfyUI Version: v0.3.5-2-g497db62
  • Arguments: ComfyUI\main.py --windows-standalone-build
  • OS: nt
  • Python Version: 3.12.7 (tags/v3.12.7:0b05ead, Oct 1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]
  • Embedded Python: true
  • PyTorch Version: 2.5.1+cu124

Devices

  • Name: cuda:0 NVIDIA GeForce RTX 3060 : cudaMallocAsync
    • Type: cuda
    • VRAM Total: 12884246528
    • VRAM Free: 11788091392
    • Torch VRAM Total: 0
    • Torch VRAM Free: 0

Logs

2025-01-03T00:02:05.248361 - [START] Security scan2025-01-03T00:02:05.248361 - 
2025-01-03T00:02:11.680609 - [DONE] Security scan2025-01-03T00:02:11.680609 - 
2025-01-03T00:03:05.240208 - Starting server

2025-01-03T00:03:05.240208 - To see the GUI go to: http://127.0.0.1:8188
2025-01-03T00:03:11.924211 - FETCH DATA from: D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json2025-01-03T00:03:11.951207 -  [DONE]2025-01-03T00:03:11.952207 - 
2025-01-03T00:03:24.934161 - got prompt
2025-01-03T00:03:28.107182 - Missing VAE keys2025-01-03T00:03:28.108184 -  2025-01-03T00:03:28.108184 - ['encoder.project_in.weight', 'encoder.project_in.bias', 'encoder.stages.0.0.conv1.conv.weight', 'encoder.stages.0.0.conv1.conv.bias', 'encoder.stages.0.0.conv2.conv.weight', 'encoder.stages.0.0.conv2.norm.weight', 'encoder.stages.0.0.conv2.norm.bias', 'encoder.stages.0.1.conv1.conv.weight', 'encoder.stages.0.1.conv1.conv.bias', 'encoder.stages.0.1.conv2.conv.weight', 'encoder.stages.0.1.conv2.norm.weight', 'encoder.stages.0.1.conv2.norm.bias', 'encoder.stages.0.2.main.weight', 'encoder.stages.0.2.main.bias', 'encoder.stages.1.0.conv1.conv.weight', 'encoder.stages.1.0.conv1.conv.bias', 'encoder.stages.1.0.conv2.conv.weight', 'encoder.stages.1.0.conv2.norm.weight', 'encoder.stages.1.0.conv2.norm.bias', 'encoder.stages.1.1.conv1.conv.weight', 'encoder.stages.1.1.conv1.conv.bias', 'encoder.stages.1.1.conv2.conv.weight', 'encoder.stages.1.1.conv2.norm.weight', 'encoder.stages.1.1.conv2.norm.bias', 'encoder.stages.1.2.main.weight', 'encoder.stages.1.2.main.bias', 'encoder.stages.2.0.conv1.conv.weight', 'encoder.stages.2.0.conv1.conv.bias', 'encoder.stages.2.0.conv2.conv.weight', 'encoder.stages.2.0.conv2.norm.weight', 'encoder.stages.2.0.conv2.norm.bias', 'encoder.stages.2.1.conv1.conv.weight', 'encoder.stages.2.1.conv1.conv.bias', 'encoder.stages.2.1.conv2.conv.weight', 'encoder.stages.2.1.conv2.norm.weight', 'encoder.stages.2.1.conv2.norm.bias', 'encoder.stages.2.2.main.weight', 'encoder.stages.2.2.main.bias', 'encoder.stages.3.0.context_module.qkv.0.weight', 'encoder.stages.3.0.context_module.aggreg.0.0.weight', 'encoder.stages.3.0.context_module.aggreg.0.1.weight', 'encoder.stages.3.0.context_module.proj.0.weight', 'encoder.stages.3.0.context_module.proj.1.weight', 'encoder.stages.3.0.context_module.proj.1.bias', 'encoder.stages.3.0.local_module.inverted_conv.conv.weight', 'encoder.stages.3.0.local_module.inverted_conv.conv.bias', 'encoder.stages.3.0.local_module.depth_conv.conv.weight', 'encoder.stages.3.0.local_module.depth_conv.conv.bias', 'encoder.stages.3.0.local_module.point_conv.conv.weight', 'encoder.stages.3.0.local_module.point_conv.norm.weight', 'encoder.stages.3.0.local_module.point_conv.norm.bias', 'encoder.stages.3.1.context_module.qkv.0.weight', 'encoder.stages.3.1.context_module.aggreg.0.0.weight', 'encoder.stages.3.1.context_module.aggreg.0.1.weight', 'encoder.stages.3.1.context_module.proj.0.weight', 'encoder.stages.3.1.context_module.proj.1.weight', 'encoder.stages.3.1.context_module.proj.1.bias', 'encoder.stages.3.1.local_module.inverted_conv.conv.weight', 'encoder.stages.3.1.local_module.inverted_conv.conv.bias', 'encoder.stages.3.1.local_module.depth_conv.conv.weight', 'encoder.stages.3.1.local_module.depth_conv.conv.bias', 'encoder.stages.3.1.local_module.point_conv.conv.weight', 'encoder.stages.3.1.local_module.point_conv.norm.weight', 'encoder.stages.3.1.local_module.point_conv.norm.bias', 'encoder.stages.3.2.context_module.qkv.0.weight', 'encoder.stages.3.2.context_module.aggreg.0.0.weight', 'encoder.stages.3.2.context_module.aggreg.0.1.weight', 'encoder.stages.3.2.context_module.proj.0.weight', 'encoder.stages.3.2.context_module.proj.1.weight', 'encoder.stages.3.2.context_module.proj.1.bias', 'encoder.stages.3.2.local_module.inverted_conv.conv.weight', 'encoder.stages.3.2.local_module.inverted_conv.conv.bias', 'encoder.stages.3.2.local_module.depth_conv.conv.weight', 'encoder.stages.3.2.local_module.depth_conv.conv.bias', 'encoder.stages.3.2.local_module.point_conv.conv.weight', 'encoder.stages.3.2.local_module.point_conv.norm.weight', 'encoder.stages.3.2.local_module.point_conv.norm.bias', 'encoder.stages.3.3.main.weight', 'encoder.stages.3.3.main.bias', 'encoder.stages.4.0.context_module.qkv.0.weight', 'encoder.stages.4.0.context_module.aggreg.0.0.weight', 'encoder.stages.4.0.context_module.aggreg.0.1.weight', 'encoder.stages.4.0.context_module.proj.0.weight', 'encoder.stages.4.0.context_module.proj.1.weight', 'encoder.stages.4.0.context_module.proj.1.bias', 'encoder.stages.4.0.local_module.inverted_conv.conv.weight', 'encoder.stages.4.0.local_module.inverted_conv.conv.bias', 'encoder.stages.4.0.local_module.depth_conv.conv.weight', 'encoder.stages.4.0.local_module.depth_conv.conv.bias', 'encoder.stages.4.0.local_module.point_conv.conv.weight', 'encoder.stages.4.0.local_module.point_conv.norm.weight', 'encoder.stages.4.0.local_module.point_conv.norm.bias', 'encoder.stages.4.1.context_module.qkv.0.weight', 'encoder.stages.4.1.context_module.aggreg.0.0.weight', 'encoder.stages.4.1.context_module.aggreg.0.1.weight', 'encoder.stages.4.1.context_module.proj.0.weight', 'encoder.stages.4.1.context_module.proj.1.weight', 'encoder.stages.4.1.context_module.proj.1.bias', 'encoder.stages.4.1.local_module.inverted_conv.conv.weight', 'encoder.stages.4.1.local_module.inverted_conv.conv.bias', 'encoder.stages.4.1.local_module.depth_conv.conv.weight', 'encoder.stages.4.1.local_module.depth_conv.conv.bias', 'encoder.stages.4.1.local_module.point_conv.conv.weight', 'encoder.stages.4.1.local_module.point_conv.norm.weight', 'encoder.stages.4.1.local_module.point_conv.norm.bias', 'encoder.stages.4.2.context_module.qkv.0.weight', 'encoder.stages.4.2.context_module.aggreg.0.0.weight', 'encoder.stages.4.2.context_module.aggreg.0.1.weight', 'encoder.stages.4.2.context_module.proj.0.weight', 'encoder.stages.4.2.context_module.proj.1.weight', 'encoder.stages.4.2.context_module.proj.1.bias', 'encoder.stages.4.2.local_module.inverted_conv.conv.weight', 'encoder.stages.4.2.local_module.inverted_conv.conv.bias', 'encoder.stages.4.2.local_module.depth_conv.conv.weight', 'encoder.stages.4.2.local_module.depth_conv.conv.bias', 'encoder.stages.4.2.local_module.point_conv.conv.weight', 'encoder.stages.4.2.local_module.point_conv.norm.weight', 'encoder.stages.4.2.local_module.point_conv.norm.bias', 'encoder.stages.4.3.main.weight', 'encoder.stages.4.3.main.bias', 'encoder.stages.5.0.context_module.qkv.0.weight', 'encoder.stages.5.0.context_module.aggreg.0.0.weight', 'encoder.stages.5.0.context_module.aggreg.0.1.weight', 'encoder.stages.5.0.context_module.proj.0.weight', 'encoder.stages.5.0.context_module.proj.1.weight', 'encoder.stages.5.0.context_module.proj.1.bias', 'encoder.stages.5.0.local_module.inverted_conv.conv.weight', 'encoder.stages.5.0.local_module.inverted_conv.conv.bias', 'encoder.stages.5.0.local_module.depth_conv.conv.weight', 'encoder.stages.5.0.local_module.depth_conv.conv.bias', 'encoder.stages.5.0.local_module.point_conv.conv.weight', 'encoder.stages.5.0.local_module.point_conv.norm.weight', 'encoder.stages.5.0.local_module.point_conv.norm.bias', 'encoder.stages.5.1.context_module.qkv.0.weight', 'encoder.stages.5.1.context_module.aggreg.0.0.weight', 'encoder.stages.5.1.context_module.aggreg.0.1.weight', 'encoder.stages.5.1.context_module.proj.0.weight', 'encoder.stages.5.1.context_module.proj.1.weight', 'encoder.stages.5.1.context_module.proj.1.bias', 'encoder.stages.5.1.local_module.inverted_conv.conv.weight', 'encoder.stages.5.1.local_module.inverted_conv.conv.bias', 'encoder.stages.5.1.local_module.depth_conv.conv.weight', 'encoder.stages.5.1.local_module.depth_conv.conv.bias', 'encoder.stages.5.1.local_module.point_conv.conv.weight', 'encoder.stages.5.1.local_module.point_conv.norm.weight', 'encoder.stages.5.1.local_module.point_conv.norm.bias', 'encoder.stages.5.2.context_module.qkv.0.weight', 'encoder.stages.5.2.context_module.aggreg.0.0.weight', 'encoder.stages.5.2.context_module.aggreg.0.1.weight', 'encoder.stages.5.2.context_module.proj.0.weight', 'encoder.stages.5.2.context_module.proj.1.weight', 'encoder.stages.5.2.context_module.proj.1.bias', 'encoder.stages.5.2.local_module.inverted_conv.conv.weight', 'encoder.stages.5.2.local_module.inverted_conv.conv.bias', 'encoder.stages.5.2.local_module.depth_conv.conv.weight', 'encoder.stages.5.2.local_module.depth_conv.conv.bias', 'encoder.stages.5.2.local_module.point_conv.conv.weight', 'encoder.stages.5.2.local_module.point_conv.norm.weight', 'encoder.stages.5.2.local_module.point_conv.norm.bias', 'encoder.project_out.main.0.conv.weight', 'encoder.project_out.main.0.conv.bias', 'decoder.project_in.main.conv.weight', 'decoder.project_in.main.conv.bias', 'decoder.stages.0.0.main.conv.weight', 'decoder.stages.0.0.main.conv.bias', 'decoder.stages.0.1.conv1.conv.weight', 'decoder.stages.0.1.conv1.conv.bias', 'decoder.stages.0.1.conv2.conv.weight', 'decoder.stages.0.1.conv2.norm.weight', 'decoder.stages.0.1.conv2.norm.bias', 'decoder.stages.0.2.conv1.conv.weight', 'decoder.stages.0.2.conv1.conv.bias', 'decoder.stages.0.2.conv2.conv.weight', 'decoder.stages.0.2.conv2.norm.weight', 'decoder.stages.0.2.conv2.norm.bias', 'decoder.stages.0.3.conv1.conv.weight', 'decoder.stages.0.3.conv1.conv.bias', 'decoder.stages.0.3.conv2.conv.weight', 'decoder.stages.0.3.conv2.norm.weight', 'decoder.stages.0.3.conv2.norm.bias', 'decoder.stages.1.0.main.conv.weight', 'decoder.stages.1.0.main.conv.bias', 'decoder.stages.1.1.conv1.conv.weight', 'decoder.stages.1.1.conv1.conv.bias', 'decoder.stages.1.1.conv2.conv.weight', 'decoder.stages.1.1.conv2.norm.weight', 'decoder.stages.1.1.conv2.norm.bias', 'decoder.stages.1.2.conv1.conv.weight', 'decoder.stages.1.2.conv1.conv.bias', 'decoder.stages.1.2.conv2.conv.weight', 'decoder.stages.1.2.conv2.norm.weight', 'decoder.stages.1.2.conv2.norm.bias', 'decoder.stages.1.3.conv1.conv.weight', 'decoder.stages.1.3.conv1.conv.bias', 'decoder.stages.1.3.conv2.conv.weight', 'decoder.stages.1.3.conv2.norm.weight', 'decoder.stages.1.3.conv2.norm.bias', 'decoder.stages.2.0.main.conv.weight', 'decoder.stages.2.0.main.conv.bias', 'decoder.stages.2.1.conv1.conv.weight', 'decoder.stages.2.1.conv1.conv.bias', 'decoder.stages.2.1.conv2.conv.weight', 'decoder.stages.2.1.conv2.norm.weight', 'decoder.stages.2.1.conv2.norm.bias', 'decoder.stages.2.2.conv1.conv.weight', 'decoder.stages.2.2.conv1.conv.bias', 'decoder.stages.2.2.conv2.conv.weight', 'decoder.stages.2.2.conv2.norm.weight', 'decoder.stages.2.2.conv2.norm.bias', 'decoder.stages.2.3.conv1.conv.weight', 'decoder.stages.2.3.conv1.conv.bias', 'decoder.stages.2.3.conv2.conv.weight', 'decoder.stages.2.3.conv2.norm.weight', 'decoder.stages.2.3.conv2.norm.bias', 'decoder.stages.3.0.main.conv.weight', 'decoder.stages.3.0.main.conv.bias', 'decoder.stages.3.1.context_module.qkv.0.weight', 'decoder.stages.3.1.context_module.aggreg.0.0.weight', 'decoder.stages.3.1.context_module.aggreg.0.1.weight', 'decoder.stages.3.1.context_module.proj.0.weight', 'decoder.stages.3.1.context_module.proj.1.weight', 'decoder.stages.3.1.context_module.proj.1.bias', 'decoder.stages.3.1.local_module.inverted_conv.conv.weight', 'decoder.stages.3.1.local_module.inverted_conv.conv.bias', 'decoder.stages.3.1.local_module.depth_conv.conv.weight', 'decoder.stages.3.1.local_module.depth_conv.conv.bias', 'decoder.stages.3.1.local_module.point_conv.conv.weight', 'decoder.stages.3.1.local_module.point_conv.norm.weight', 'decoder.stages.3.1.local_module.point_conv.norm.bias', 'decoder.stages.3.2.context_module.qkv.0.weight', 'decoder.stages.3.2.context_module.aggreg.0.0.weight', 'decoder.stages.3.2.context_module.aggreg.0.1.weight', 'decoder.stages.3.2.context_module.proj.0.weight', 'decoder.stages.3.2.context_module.proj.1.weight', 'decoder.stages.3.2.context_module.proj.1.bias', 'decoder.stages.3.2.local_module.inverted_conv.conv.weight', 'decoder.stages.3.2.local_module.inverted_conv.conv.bias', 'decoder.stages.3.2.local_module.depth_conv.conv.weight', 'decoder.stages.3.2.local_module.depth_conv.conv.bias', 'decoder.stages.3.2.local_module.point_conv.conv.weight', 'decoder.stages.3.2.local_module.point_conv.norm.weight', 'decoder.stages.3.2.local_module.point_conv.norm.bias', 'decoder.stages.3.3.context_module.qkv.0.weight', 'decoder.stages.3.3.context_module.aggreg.0.0.weight', 'decoder.stages.3.3.context_module.aggreg.0.1.weight', 'decoder.stages.3.3.context_module.proj.0.weight', 'decoder.stages.3.3.context_module.proj.1.weight', 'decoder.stages.3.3.context_module.proj.1.bias', 'decoder.stages.3.3.local_module.inverted_conv.conv.weight', 'decoder.stages.3.3.local_module.inverted_conv.conv.bias', 'decoder.stages.3.3.local_module.depth_conv.conv.weight', 'decoder.stages.3.3.local_module.depth_conv.conv.bias', 'decoder.stages.3.3.local_module.point_conv.conv.weight', 'decoder.stages.3.3.local_module.point_conv.norm.weight', 'decoder.stages.3.3.local_module.point_conv.norm.bias', 'decoder.stages.4.0.main.conv.weight', 'decoder.stages.4.0.main.conv.bias', 'decoder.stages.4.1.context_module.qkv.0.weight', 'decoder.stages.4.1.context_module.aggreg.0.0.weight', 'decoder.stages.4.1.context_module.aggreg.0.1.weight', 'decoder.stages.4.1.context_module.proj.0.weight', 'decoder.stages.4.1.context_module.proj.1.weight', 'decoder.stages.4.1.context_module.proj.1.bias', 'decoder.stages.4.1.local_module.inverted_conv.conv.weight', 'decoder.stages.4.1.local_module.inverted_conv.conv.bias', 'decoder.stages.4.1.local_module.depth_conv.conv.weight', 'decoder.stages.4.1.local_module.depth_conv.conv.bias', 'decoder.stages.4.1.local_module.point_conv.conv.weight', 'decoder.stages.4.1.local_module.point_conv.norm.weight', 'decoder.stages.4.1.local_module.point_conv.norm.bias', 'decoder.stages.4.2.context_module.qkv.0.weight', 'decoder.stages.4.2.context_module.aggreg.0.0.weight', 'decoder.stages.4.2.context_module.aggreg.0.1.weight', 'decoder.stages.4.2.context_module.proj.0.weight', 'decoder.stages.4.2.context_module.proj.1.weight', 'decoder.stages.4.2.context_module.proj.1.bias', 'decoder.stages.4.2.local_module.inverted_conv.conv.weight', 'decoder.stages.4.2.local_module.inverted_conv.conv.bias', 'decoder.stages.4.2.local_module.depth_conv.conv.weight', 'decoder.stages.4.2.local_module.depth_conv.conv.bias', 'decoder.stages.4.2.local_module.point_conv.conv.weight', 'decoder.stages.4.2.local_module.point_conv.norm.weight', 'decoder.stages.4.2.local_module.point_conv.norm.bias', 'decoder.stages.4.3.context_module.qkv.0.weight', 'decoder.stages.4.3.context_module.aggreg.0.0.weight', 'decoder.stages.4.3.context_module.aggreg.0.1.weight', 'decoder.stages.4.3.context_module.proj.0.weight', 'decoder.stages.4.3.context_module.proj.1.weight', 'decoder.stages.4.3.context_module.proj.1.bias', 'decoder.stages.4.3.local_module.inverted_conv.conv.weight', 'decoder.stages.4.3.local_module.inverted_conv.conv.bias', 'decoder.stages.4.3.local_module.depth_conv.conv.weight', 'decoder.stages.4.3.local_module.depth_conv.conv.bias', 'decoder.stages.4.3.local_module.point_conv.conv.weight', 'decoder.stages.4.3.local_module.point_conv.norm.weight', 'decoder.stages.4.3.local_module.point_conv.norm.bias', 'decoder.stages.5.0.context_module.qkv.0.weight', 'decoder.stages.5.0.context_module.aggreg.0.0.weight', 'decoder.stages.5.0.context_module.aggreg.0.1.weight', 'decoder.stages.5.0.context_module.proj.0.weight', 'decoder.stages.5.0.context_module.proj.1.weight', 'decoder.stages.5.0.context_module.proj.1.bias', 'decoder.stages.5.0.local_module.inverted_conv.conv.weight', 'decoder.stages.5.0.local_module.inverted_conv.conv.bias', 'decoder.stages.5.0.local_module.depth_conv.conv.weight', 'decoder.stages.5.0.local_module.depth_conv.conv.bias', 'decoder.stages.5.0.local_module.point_conv.conv.weight', 'decoder.stages.5.0.local_module.point_conv.norm.weight', 'decoder.stages.5.0.local_module.point_conv.norm.bias', 'decoder.stages.5.1.context_module.qkv.0.weight', 'decoder.stages.5.1.context_module.aggreg.0.0.weight', 'decoder.stages.5.1.context_module.aggreg.0.1.weight', 'decoder.stages.5.1.context_module.proj.0.weight', 'decoder.stages.5.1.context_module.proj.1.weight', 'decoder.stages.5.1.context_module.proj.1.bias', 'decoder.stages.5.1.local_module.inverted_conv.conv.weight', 'decoder.stages.5.1.local_module.inverted_conv.conv.bias', 'decoder.stages.5.1.local_module.depth_conv.conv.weight', 'decoder.stages.5.1.local_module.depth_conv.conv.bias', 'decoder.stages.5.1.local_module.point_conv.conv.weight', 'decoder.stages.5.1.local_module.point_conv.norm.weight', 'decoder.stages.5.1.local_module.point_conv.norm.bias', 'decoder.stages.5.2.context_module.qkv.0.weight', 'decoder.stages.5.2.context_module.aggreg.0.0.weight', 'decoder.stages.5.2.context_module.aggreg.0.1.weight', 'decoder.stages.5.2.context_module.proj.0.weight', 'decoder.stages.5.2.context_module.proj.1.weight', 'decoder.stages.5.2.context_module.proj.1.bias', 'decoder.stages.5.2.local_module.inverted_conv.conv.weight', 'decoder.stages.5.2.local_module.inverted_conv.conv.bias', 'decoder.stages.5.2.local_module.depth_conv.conv.weight', 'decoder.stages.5.2.local_module.depth_conv.conv.bias', 'decoder.stages.5.2.local_module.point_conv.conv.weight', 'decoder.stages.5.2.local_module.point_conv.norm.weight', 'decoder.stages.5.2.local_module.point_conv.norm.bias', 'decoder.project_out.0.weight', 'decoder.project_out.0.bias', 'decoder.project_out.2.conv.weight', 'decoder.project_out.2.conv.bias']2025-01-03T00:03:28.111183 - 
2025-01-03T00:03:28.111183 - Leftover VAE keys2025-01-03T00:03:28.111183 -  2025-01-03T00:03:28.111183 - ['encoder.conv_in.bias', 'encoder.conv_in.weight', 'encoder.conv_out.bias', 'encoder.conv_out.weight', 'encoder.down_blocks.0.0.conv1.bias', 'encoder.down_blocks.0.0.conv1.weight', 'encoder.down_blocks.0.0.conv2.weight', 'encoder.down_blocks.0.0.norm.bias', 'encoder.down_blocks.0.0.norm.weight', 'encoder.down_blocks.0.1.conv1.bias', 'encoder.down_blocks.0.1.conv1.weight', 'encoder.down_blocks.0.1.conv2.weight', 'encoder.down_blocks.0.1.norm.bias', 'encoder.down_blocks.0.1.norm.weight', 'encoder.down_blocks.0.2.conv.bias', 'encoder.down_blocks.0.2.conv.weight', 'encoder.down_blocks.1.0.conv1.bias', 'encoder.down_blocks.1.0.conv1.weight', 'encoder.down_blocks.1.0.conv2.weight', 'encoder.down_blocks.1.0.norm.bias', 'encoder.down_blocks.1.0.norm.weight', 'encoder.down_blocks.1.1.conv1.bias', 'encoder.down_blocks.1.1.conv1.weight', 'encoder.down_blocks.1.1.conv2.weight', 'encoder.down_blocks.1.1.norm.bias', 'encoder.down_blocks.1.1.norm.weight', 'encoder.down_blocks.1.2.conv.bias', 'encoder.down_blocks.1.2.conv.weight', 'encoder.down_blocks.2.0.conv1.bias', 'encoder.down_blocks.2.0.conv1.weight', 'encoder.down_blocks.2.0.conv2.weight', 'encoder.down_blocks.2.0.norm.bias', 'encoder.down_blocks.2.0.norm.weight', 'encoder.down_blocks.2.1.conv1.bias', 'encoder.down_blocks.2.1.conv1.weight', 'encoder.down_blocks.2.1.conv2.weight', 'encoder.down_blocks.2.1.norm.bias', 'encoder.down_blocks.2.1.norm.weight', 'encoder.down_blocks.2.2.conv.bias', 'encoder.down_blocks.2.2.conv.weight', 'encoder.down_blocks.3.0.attn.norm_out.bias', 'encoder.down_blocks.3.0.attn.norm_out.weight', 'encoder.down_blocks.3.0.attn.to_k.weight', 'encoder.down_blocks.3.0.attn.to_out.weight', 'encoder.down_blocks.3.0.attn.to_q.weight', 'encoder.down_blocks.3.0.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.3.0.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.3.0.attn.to_v.weight', 'encoder.down_blocks.3.0.conv_out.conv_depth.bias', 'encoder.down_blocks.3.0.conv_out.conv_depth.weight', 'encoder.down_blocks.3.0.conv_out.conv_inverted.bias', 'encoder.down_blocks.3.0.conv_out.conv_inverted.weight', 'encoder.down_blocks.3.0.conv_out.conv_point.weight', 'encoder.down_blocks.3.0.conv_out.norm.bias', 'encoder.down_blocks.3.0.conv_out.norm.weight', 'encoder.down_blocks.3.1.attn.norm_out.bias', 'encoder.down_blocks.3.1.attn.norm_out.weight', 'encoder.down_blocks.3.1.attn.to_k.weight', 'encoder.down_blocks.3.1.attn.to_out.weight', 'encoder.down_blocks.3.1.attn.to_q.weight', 'encoder.down_blocks.3.1.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.3.1.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.3.1.attn.to_v.weight', 'encoder.down_blocks.3.1.conv_out.conv_depth.bias', 'encoder.down_blocks.3.1.conv_out.conv_depth.weight', 'encoder.down_blocks.3.1.conv_out.conv_inverted.bias', 'encoder.down_blocks.3.1.conv_out.conv_inverted.weight', 'encoder.down_blocks.3.1.conv_out.conv_point.weight', 'encoder.down_blocks.3.1.conv_out.norm.bias', 'encoder.down_blocks.3.1.conv_out.norm.weight', 'encoder.down_blocks.3.2.attn.norm_out.bias', 'encoder.down_blocks.3.2.attn.norm_out.weight', 'encoder.down_blocks.3.2.attn.to_k.weight', 'encoder.down_blocks.3.2.attn.to_out.weight', 'encoder.down_blocks.3.2.attn.to_q.weight', 'encoder.down_blocks.3.2.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.3.2.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.3.2.attn.to_v.weight', 'encoder.down_blocks.3.2.conv_out.conv_depth.bias', 'encoder.down_blocks.3.2.conv_out.conv_depth.weight', 'encoder.down_blocks.3.2.conv_out.conv_inverted.bias', 'encoder.down_blocks.3.2.conv_out.conv_inverted.weight', 'encoder.down_blocks.3.2.conv_out.conv_point.weight', 'encoder.down_blocks.3.2.conv_out.norm.bias', 'encoder.down_blocks.3.2.conv_out.norm.weight', 'encoder.down_blocks.3.3.conv.bias', 'encoder.down_blocks.3.3.conv.weight', 'encoder.down_blocks.4.0.attn.norm_out.bias', 'encoder.down_blocks.4.0.attn.norm_out.weight', 'encoder.down_blocks.4.0.attn.to_k.weight', 'encoder.down_blocks.4.0.attn.to_out.weight', 'encoder.down_blocks.4.0.attn.to_q.weight', 'encoder.down_blocks.4.0.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.4.0.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.4.0.attn.to_v.weight', 'encoder.down_blocks.4.0.conv_out.conv_depth.bias', 'encoder.down_blocks.4.0.conv_out.conv_depth.weight', 'encoder.down_blocks.4.0.conv_out.conv_inverted.bias', 'encoder.down_blocks.4.0.conv_out.conv_inverted.weight', 'encoder.down_blocks.4.0.conv_out.conv_point.weight', 'encoder.down_blocks.4.0.conv_out.norm.bias', 'encoder.down_blocks.4.0.conv_out.norm.weight', 'encoder.down_blocks.4.1.attn.norm_out.bias', 'encoder.down_blocks.4.1.attn.norm_out.weight', 'encoder.down_blocks.4.1.attn.to_k.weight', 'encoder.down_blocks.4.1.attn.to_out.weight', 'encoder.down_blocks.4.1.attn.to_q.weight', 'encoder.down_blocks.4.1.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.4.1.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.4.1.attn.to_v.weight', 'encoder.down_blocks.4.1.conv_out.conv_depth.bias', 'encoder.down_blocks.4.1.conv_out.conv_depth.weight', 'encoder.down_blocks.4.1.conv_out.conv_inverted.bias', 'encoder.down_blocks.4.1.conv_out.conv_inverted.weight', 'encoder.down_blocks.4.1.conv_out.conv_point.weight', 'encoder.down_blocks.4.1.conv_out.norm.bias', 'encoder.down_blocks.4.1.conv_out.norm.weight', 'encoder.down_blocks.4.2.attn.norm_out.bias', 'encoder.down_blocks.4.2.attn.norm_out.weight', 'encoder.down_blocks.4.2.attn.to_k.weight', 'encoder.down_blocks.4.2.attn.to_out.weight', 'encoder.down_blocks.4.2.attn.to_q.weight', 'encoder.down_blocks.4.2.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.4.2.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.4.2.attn.to_v.weight', 'encoder.down_blocks.4.2.conv_out.conv_depth.bias', 'encoder.down_blocks.4.2.conv_out.conv_depth.weight', 'encoder.down_blocks.4.2.conv_out.conv_inverted.bias', 'encoder.down_blocks.4.2.conv_out.conv_inverted.weight', 'encoder.down_blocks.4.2.conv_out.conv_point.weight', 'encoder.down_blocks.4.2.conv_out.norm.bias', 'encoder.down_blocks.4.2.conv_out.norm.weight', 'encoder.down_blocks.4.3.conv.bias', 'encoder.down_blocks.4.3.conv.weight', 'encoder.down_blocks.5.0.attn.norm_out.bias', 'encoder.down_blocks.5.0.attn.norm_out.weight', 'encoder.down_blocks.5.0.attn.to_k.weight', 'encoder.down_blocks.5.0.attn.to_out.weight', 'encoder.down_blocks.5.0.attn.to_q.weight', 'encoder.down_blocks.5.0.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.5.0.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.5.0.attn.to_v.weight', 'encoder.down_blocks.5.0.conv_out.conv_depth.bias', 'encoder.down_blocks.5.0.conv_out.conv_depth.weight', 'encoder.down_blocks.5.0.conv_out.conv_inverted.bias', 'encoder.down_blocks.5.0.conv_out.conv_inverted.weight', 'encoder.down_blocks.5.0.conv_out.conv_point.weight', 'encoder.down_blocks.5.0.conv_out.norm.bias', 'encoder.down_blocks.5.0.conv_out.norm.weight', 'encoder.down_blocks.5.1.attn.norm_out.bias', 'encoder.down_blocks.5.1.attn.norm_out.weight', 'encoder.down_blocks.5.1.attn.to_k.weight', 'encoder.down_blocks.5.1.attn.to_out.weight', 'encoder.down_blocks.5.1.attn.to_q.weight', 'encoder.down_blocks.5.1.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.5.1.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.5.1.attn.to_v.weight', 'encoder.down_blocks.5.1.conv_out.conv_depth.bias', 'encoder.down_blocks.5.1.conv_out.conv_depth.weight', 'encoder.down_blocks.5.1.conv_out.conv_inverted.bias', 'encoder.down_blocks.5.1.conv_out.conv_inverted.weight', 'encoder.down_blocks.5.1.conv_out.conv_point.weight', 'encoder.down_blocks.5.1.conv_out.norm.bias', 'encoder.down_blocks.5.1.conv_out.norm.weight', 'encoder.down_blocks.5.2.attn.norm_out.bias', 'encoder.down_blocks.5.2.attn.norm_out.weight', 'encoder.down_blocks.5.2.attn.to_k.weight', 'encoder.down_blocks.5.2.attn.to_out.weight', 'encoder.down_blocks.5.2.attn.to_q.weight', 'encoder.down_blocks.5.2.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.5.2.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.5.2.attn.to_v.weight', 'encoder.down_blocks.5.2.conv_out.conv_depth.bias', 'encoder.down_blocks.5.2.conv_out.conv_depth.weight', 'encoder.down_blocks.5.2.conv_out.conv_inverted.bias', 'encoder.down_blocks.5.2.conv_out.conv_inverted.weight', 'encoder.down_blocks.5.2.conv_out.conv_point.weight', 'encoder.down_blocks.5.2.conv_out.norm.bias', 'encoder.down_blocks.5.2.conv_out.norm.weight', 'decoder.conv_in.bias', 'decoder.conv_in.weight', 'decoder.conv_out.bias', 'decoder.conv_out.weight', 'decoder.norm_out.bias', 'decoder.norm_out.weight', 'decoder.up_blocks.0.0.conv.bias', 'decoder.up_blocks.0.0.conv.weight', 'decoder.up_blocks.0.1.conv1.bias', 'decoder.up_blocks.0.1.conv1.weight', 'decoder.up_blocks.0.1.conv2.weight', 'decoder.up_blocks.0.1.norm.bias', 'decoder.up_blocks.0.1.norm.weight', 'decoder.up_blocks.0.2.conv1.bias', 'decoder.up_blocks.0.2.conv1.weight', 'decoder.up_blocks.0.2.conv2.weight', 'decoder.up_blocks.0.2.norm.bias', 'decoder.up_blocks.0.2.norm.weight', 'decoder.up_blocks.0.3.conv1.bias', 'decoder.up_blocks.0.3.conv1.weight', 'decoder.up_blocks.0.3.conv2.weight', 'decoder.up_blocks.0.3.norm.bias', 'decoder.up_blocks.0.3.norm.weight', 'decoder.up_blocks.1.0.conv.bias', 'decoder.up_blocks.1.0.conv.weight', 'decoder.up_blocks.1.1.conv1.bias', 'decoder.up_blocks.1.1.conv1.weight', 'decoder.up_blocks.1.1.conv2.weight', 'decoder.up_blocks.1.1.norm.bias', 'decoder.up_blocks.1.1.norm.weight', 'decoder.up_blocks.1.2.conv1.bias', 'decoder.up_blocks.1.2.conv1.weight', 'decoder.up_blocks.1.2.conv2.weight', 'decoder.up_blocks.1.2.norm.bias', 'decoder.up_blocks.1.2.norm.weight', 'decoder.up_blocks.1.3.conv1.bias', 'decoder.up_blocks.1.3.conv1.weight', 'decoder.up_blocks.1.3.conv2.weight', 'decoder.up_blocks.1.3.norm.bias', 'decoder.up_blocks.1.3.norm.weight', 'decoder.up_blocks.2.0.conv.bias', 'decoder.up_blocks.2.0.conv.weight', 'decoder.up_blocks.2.1.conv1.bias', 'decoder.up_blocks.2.1.conv1.weight', 'decoder.up_blocks.2.1.conv2.weight', 'decoder.up_blocks.2.1.norm.bias', 'decoder.up_blocks.2.1.norm.weight', 'decoder.up_blocks.2.2.conv1.bias', 'decoder.up_blocks.2.2.conv1.weight', 'decoder.up_blocks.2.2.conv2.weight', 'decoder.up_blocks.2.2.norm.bias', 'decoder.up_blocks.2.2.norm.weight', 'decoder.up_blocks.2.3.conv1.bias', 'decoder.up_blocks.2.3.conv1.weight', 'decoder.up_blocks.2.3.conv2.weight', 'decoder.up_blocks.2.3.norm.bias', 'decoder.up_blocks.2.3.norm.weight', 'decoder.up_blocks.3.0.conv.bias', 'decoder.up_blocks.3.0.conv.weight', 'decoder.up_blocks.3.1.attn.norm_out.bias', 'decoder.up_blocks.3.1.attn.norm_out.weight', 'decoder.up_blocks.3.1.attn.to_k.weight', 'decoder.up_blocks.3.1.attn.to_out.weight', 'decoder.up_blocks.3.1.attn.to_q.weight', 'decoder.up_blocks.3.1.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.3.1.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.3.1.attn.to_v.weight', 'decoder.up_blocks.3.1.conv_out.conv_depth.bias', 'decoder.up_blocks.3.1.conv_out.conv_depth.weight', 'decoder.up_blocks.3.1.conv_out.conv_inverted.bias', 'decoder.up_blocks.3.1.conv_out.conv_inverted.weight', 'decoder.up_blocks.3.1.conv_out.conv_point.weight', 'decoder.up_blocks.3.1.conv_out.norm.bias', 'decoder.up_blocks.3.1.conv_out.norm.weight', 'decoder.up_blocks.3.2.attn.norm_out.bias', 'decoder.up_blocks.3.2.attn.norm_out.weight', 'decoder.up_blocks.3.2.attn.to_k.weight', 'decoder.up_blocks.3.2.attn.to_out.weight', 'decoder.up_blocks.3.2.attn.to_q.weight', 'decoder.up_blocks.3.2.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.3.2.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.3.2.attn.to_v.weight', 'decoder.up_blocks.3.2.conv_out.conv_depth.bias', 'decoder.up_blocks.3.2.conv_out.conv_depth.weight', 'decoder.up_blocks.3.2.conv_out.conv_inverted.bias', 'decoder.up_blocks.3.2.conv_out.conv_inverted.weight', 'decoder.up_blocks.3.2.conv_out.conv_point.weight', 'decoder.up_blocks.3.2.conv_out.norm.bias', 'decoder.up_blocks.3.2.conv_out.norm.weight', 'decoder.up_blocks.3.3.attn.norm_out.bias', 'decoder.up_blocks.3.3.attn.norm_out.weight', 'decoder.up_blocks.3.3.attn.to_k.weight', 'decoder.up_blocks.3.3.attn.to_out.weight', 'decoder.up_blocks.3.3.attn.to_q.weight', 'decoder.up_blocks.3.3.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.3.3.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.3.3.attn.to_v.weight', 'decoder.up_blocks.3.3.conv_out.conv_depth.bias', 'decoder.up_blocks.3.3.conv_out.conv_depth.weight', 'decoder.up_blocks.3.3.conv_out.conv_inverted.bias', 'decoder.up_blocks.3.3.conv_out.conv_inverted.weight', 'decoder.up_blocks.3.3.conv_out.conv_point.weight', 'decoder.up_blocks.3.3.conv_out.norm.bias', 'decoder.up_blocks.3.3.conv_out.norm.weight', 'decoder.up_blocks.4.0.conv.bias', 'decoder.up_blocks.4.0.conv.weight', 'decoder.up_blocks.4.1.attn.norm_out.bias', 'decoder.up_blocks.4.1.attn.norm_out.weight', 'decoder.up_blocks.4.1.attn.to_k.weight', 'decoder.up_blocks.4.1.attn.to_out.weight', 'decoder.up_blocks.4.1.attn.to_q.weight', 'decoder.up_blocks.4.1.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.4.1.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.4.1.attn.to_v.weight', 'decoder.up_blocks.4.1.conv_out.conv_depth.bias', 'decoder.up_blocks.4.1.conv_out.conv_depth.weight', 'decoder.up_blocks.4.1.conv_out.conv_inverted.bias', 'decoder.up_blocks.4.1.conv_out.conv_inverted.weight', 'decoder.up_blocks.4.1.conv_out.conv_point.weight', 'decoder.up_blocks.4.1.conv_out.norm.bias', 'decoder.up_blocks.4.1.conv_out.norm.weight', 'decoder.up_blocks.4.2.attn.norm_out.bias', 'decoder.up_blocks.4.2.attn.norm_out.weight', 'decoder.up_blocks.4.2.attn.to_k.weight', 'decoder.up_blocks.4.2.attn.to_out.weight', 'decoder.up_blocks.4.2.attn.to_q.weight', 'decoder.up_blocks.4.2.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.4.2.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.4.2.attn.to_v.weight', 'decoder.up_blocks.4.2.conv_out.conv_depth.bias', 'decoder.up_blocks.4.2.conv_out.conv_depth.weight', 'decoder.up_blocks.4.2.conv_out.conv_inverted.bias', 'decoder.up_blocks.4.2.conv_out.conv_inverted.weight', 'decoder.up_blocks.4.2.conv_out.conv_point.weight', 'decoder.up_blocks.4.2.conv_out.norm.bias', 'decoder.up_blocks.4.2.conv_out.norm.weight', 'decoder.up_blocks.4.3.attn.norm_out.bias', 'decoder.up_blocks.4.3.attn.norm_out.weight', 'decoder.up_blocks.4.3.attn.to_k.weight', 'decoder.up_blocks.4.3.attn.to_out.weight', 'decoder.up_blocks.4.3.attn.to_q.weight', 'decoder.up_blocks.4.3.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.4.3.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.4.3.attn.to_v.weight', 'decoder.up_blocks.4.3.conv_out.conv_depth.bias', 'decoder.up_blocks.4.3.conv_out.conv_depth.weight', 'decoder.up_blocks.4.3.conv_out.conv_inverted.bias', 'decoder.up_blocks.4.3.conv_out.conv_inverted.weight', 'decoder.up_blocks.4.3.conv_out.conv_point.weight', 'decoder.up_blocks.4.3.conv_out.norm.bias', 'decoder.up_blocks.4.3.conv_out.norm.weight', 'decoder.up_blocks.5.0.attn.norm_out.bias', 'decoder.up_blocks.5.0.attn.norm_out.weight', 'decoder.up_blocks.5.0.attn.to_k.weight', 'decoder.up_blocks.5.0.attn.to_out.weight', 'decoder.up_blocks.5.0.attn.to_q.weight', 'decoder.up_blocks.5.0.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.5.0.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.5.0.attn.to_v.weight', 'decoder.up_blocks.5.0.conv_out.conv_depth.bias', 'decoder.up_blocks.5.0.conv_out.conv_depth.weight', 'decoder.up_blocks.5.0.conv_out.conv_inverted.bias', 'decoder.up_blocks.5.0.conv_out.conv_inverted.weight', 'decoder.up_blocks.5.0.conv_out.conv_point.weight', 'decoder.up_blocks.5.0.conv_out.norm.bias', 'decoder.up_blocks.5.0.conv_out.norm.weight', 'decoder.up_blocks.5.1.attn.norm_out.bias', 'decoder.up_blocks.5.1.attn.norm_out.weight', 'decoder.up_blocks.5.1.attn.to_k.weight', 'decoder.up_blocks.5.1.attn.to_out.weight', 'decoder.up_blocks.5.1.attn.to_q.weight', 'decoder.up_blocks.5.1.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.5.1.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.5.1.attn.to_v.weight', 'decoder.up_blocks.5.1.conv_out.conv_depth.bias', 'decoder.up_blocks.5.1.conv_out.conv_depth.weight', 'decoder.up_blocks.5.1.conv_out.conv_inverted.bias', 'decoder.up_blocks.5.1.conv_out.conv_inverted.weight', 'decoder.up_blocks.5.1.conv_out.conv_point.weight', 'decoder.up_blocks.5.1.conv_out.norm.bias', 'decoder.up_blocks.5.1.conv_out.norm.weight', 'decoder.up_blocks.5.2.attn.norm_out.bias', 'decoder.up_blocks.5.2.attn.norm_out.weight', 'decoder.up_blocks.5.2.attn.to_k.weight', 'decoder.up_blocks.5.2.attn.to_out.weight', 'decoder.up_blocks.5.2.attn.to_q.weight', 'decoder.up_blocks.5.2.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.5.2.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.5.2.attn.to_v.weight', 'decoder.up_blocks.5.2.conv_out.conv_depth.bias', 'decoder.up_blocks.5.2.conv_out.conv_depth.weight', 'decoder.up_blocks.5.2.conv_out.conv_inverted.bias', 'decoder.up_blocks.5.2.conv_out.conv_inverted.weight', 'decoder.up_blocks.5.2.conv_out.conv_point.weight', 'decoder.up_blocks.5.2.conv_out.norm.bias', 'decoder.up_blocks.5.2.conv_out.norm.weight']2025-01-03T00:03:28.114182 - 
2025-01-03T00:03:29.402325 - !!! Exception during processing !!! The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:03:29.405327 - Traceback (most recent call last):
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 945, in from_pretrained
    config_class = CONFIG_MAPPING[config_dict["model_type"]]
                   ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 647, in __getitem__
    raise KeyError(key)
KeyError: 'gemma2'

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
    text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
    config, kwargs = AutoConfig.from_pretrained(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
    raise ValueError(
ValueError: The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

2025-01-03T00:03:29.408325 - Prompt executed in 4.47 seconds
2025-01-03T00:04:14.853241 - got prompt
2025-01-03T00:04:15.726969 - !!! Exception during processing !!! The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:04:15.729969 - Traceback (most recent call last):
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 945, in from_pretrained
    config_class = CONFIG_MAPPING[config_dict["model_type"]]
                   ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 647, in __getitem__
    raise KeyError(key)
KeyError: 'gemma2'

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
    text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
    config, kwargs = AutoConfig.from_pretrained(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
    raise ValueError(
ValueError: The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

2025-01-03T00:04:15.731969 - Prompt executed in 0.88 seconds
2025-01-03T00:08:55.153890 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-01-03T00:08:55.624720 -  [DONE]2025-01-03T00:08:55.624720 - 
2025-01-03T00:09:12.246439 - 
Fetching done.2025-01-03T00:09:12.246439 - 
2025-01-03T00:14:45.570659 - got prompt
2025-01-03T00:14:56.540188 - !!! Exception during processing !!! The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:14:56.543177 - Traceback (most recent call last):
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 945, in from_pretrained
    config_class = CONFIG_MAPPING[config_dict["model_type"]]
                   ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 647, in __getitem__
    raise KeyError(key)
KeyError: 'gemma2'

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
    text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
    config, kwargs = AutoConfig.from_pretrained(
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
    raise ValueError(
ValueError: The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

2025-01-03T00:14:56.545184 - Prompt executed in 10.97 seconds

Attached Workflow

Please make sure that workflow does not contain any sensitive information such as API keys or passwords.

{"last_node_id":20,"last_link_id":33,"nodes":[{"id":1,"type":"SanaTextEncode","pos":[34.06884765625,-130.40341186523438],"size":[539.967529296875,255.81317138671875],"flags":{},"order":6,"mode":0,"inputs":[{"name":"GEMMA","type":"GEMMA","link":8}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[12,33],"slot_index":0}],"properties":{"Node name for S&R":"SanaTextEncode"},"widgets_values":["A woman with long brown hair and blue eyes, wearing a pink dress and standing in front of a blurred background of a city street."],"color":"#232","bgcolor":"#353"},{"id":2,"type":"SanaResolutionCond","pos":[605.7742919921875,-75.0198745727539],"size":[301.3979187011719,83.60631561279297],"flags":{},"order":8,"mode":0,"inputs":[{"name":"cond","type":"CONDITIONING","link":12},{"name":"width","type":"INT","link":2,"widget":{"name":"width"}},{"name":"height","type":"INT","link":3,"widget":{"name":"height"}}],"outputs":[{"name":"cond","type":"CONDITIONING","links":[13],"slot_index":0}],"properties":{"Node name for S&R":"SanaResolutionCond"},"widgets_values":[1024,1024]},{"id":4,"type":"SanaResolutionSelect","pos":[-334.3977355957031,317.1134338378906],"size":[315,102],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"width","type":"INT","links":[2,4],"slot_index":0},{"name":"height","type":"INT","links":[3,5],"slot_index":1}],"properties":{"Node name for S&R":"SanaResolutionSelect"},"widgets_values":["SanaMS_1600M_P1_D20","1.38"]},{"id":5,"type":"EmptySanaLatentImage","pos":[105.96505737304688,440.70098876953125],"size":[315,106],"flags":{},"order":5,"mode":0,"inputs":[{"name":"width","type":"INT","link":4,"widget":{"name":"width"}},{"name":"height","type":"INT","link":5,"widget":{"name":"height"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[20],"slot_index":0}],"properties":{"Node name for S&R":"EmptySanaLatentImage"},"widgets_values":[512,512,2]},{"id":6,"type":"KSampler","pos":[602.260009765625,73.22354125976562],"size":[315,262],"flags":{},"order":9,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":1},{"name":"positive","type":"CONDITIONING","link":13},{"name":"negative","type":"CONDITIONING","link":11},{"name":"latent_image","type":"LATENT","link":20},{"name":"seed","type":"INT","link":31,"widget":{"name":"seed"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[14,21],"slot_index":0}],"properties":{"Node name for S&R":"KSampler"},"widgets_values":[936979341784666,"randomize",30,3,"euler","normal",1]},{"id":9,"type":"SanaTextEncode","pos":[64.97035217285156,183.81201171875],"size":[449.4744873046875,200],"flags":{},"order":7,"mode":0,"inputs":[{"name":"GEMMA","type":"GEMMA","link":10}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[11,32],"slot_index":0}],"properties":{"Node name for S&R":"SanaTextEncode"},"widgets_values":["blurry, low quality, 3D , cartnoon , extra hands, extra fingers "],"color":"#322","bgcolor":"#533"},{"id":10,"type":"VAEDecode","pos":[699.19384765625,395.64361572265625],"size":[210,46],"flags":{},"order":10,"mode":0,"inputs":[{"name":"samples","type":"LATENT","link":14},{"name":"vae","type":"VAE","link":15}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[16],"slot_index":0}],"properties":{"Node name for S&R":"VAEDecode"},"widgets_values":[]},{"id":11,"type":"ExtraVAELoader","pos":[-416.8797302246094,501.91510009765625],"size":[398.8718566894531,106],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"VAE","type":"VAE","links":[15,23],"slot_index":0}],"properties":{"Node name for S&R":"ExtraVAELoader"},"widgets_values":["Sana_1600M_1024px_vae.safetensors","dcae-f32c32-sana-1.0","FP16"]},{"id":12,"type":"PreviewImage","pos":[1007.628662109375,-122.39791870117188],"size":[1289.3055419921875,936.1663208007812],"flags":{},"order":12,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":16}],"outputs":[],"properties":{"Node name for S&R":"PreviewImage"},"widgets_values":[]},{"id":15,"type":"KSampler","pos":[246.45791625976562,1191.375732421875],"size":[315,262],"flags":{"collapsed":true},"order":11,"mode":2,"inputs":[{"name":"model","type":"MODEL","link":28},{"name":"positive","type":"CONDITIONING","link":33},{"name":"negative","type":"CONDITIONING","link":32},{"name":"latent_image","type":"LATENT","link":21},{"name":"seed","type":"INT","link":30,"widget":{"name":"seed"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[22],"slot_index":0}],"properties":{"Node name for S&R":"KSampler"},"widgets_values":[60991530284536,"randomize",10,2,"euler","normal",0.3]},{"id":16,"type":"VAEDecode","pos":[346.4548034667969,1510.8746337890625],"size":[210,46],"flags":{"collapsed":true},"order":13,"mode":2,"inputs":[{"name":"samples","type":"LATENT","link":22},{"name":"vae","type":"VAE","link":23}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[29],"slot_index":0}],"properties":{"Node name for S&R":"VAEDecode"},"widgets_values":[]},{"id":19,"type":"PreviewImage","pos":[654.5088500976562,1181.4791259765625],"size":[1271.1556396484375,939.3509521484375],"flags":{"collapsed":true},"order":14,"mode":2,"inputs":[{"name":"images","type":"IMAGE","link":29}],"outputs":[],"properties":{"Node name for S&R":"PreviewImage"},"widgets_values":[]},{"id":20,"type":"CR Seed","pos":[106.5981674194336,632.4039306640625],"size":[315,102],"flags":{},"order":2,"mode":0,"inputs":[],"outputs":[{"name":"seed","type":"INT","links":[30,31],"slot_index":0},{"name":"show_help","type":"STRING","links":null}],"properties":{"Node name for S&R":"CR Seed"},"widgets_values":[255093259020808,"randomize"]},{"id":3,"type":"SanaCheckpointLoader","pos":[-364.4580993652344,-42.886985778808594],"size":[330.63165283203125,82],"flags":{},"order":3,"mode":0,"inputs":[],"outputs":[{"name":"model","type":"MODEL","links":[1,28],"slot_index":0}],"properties":{"Node name for S&R":"SanaCheckpointLoader"},"widgets_values":["Sana_1600M_1024px.pth","SanaMS_1600M_P1_D20"]},{"id":8,"type":"GemmaLoader","pos":[-363.2308044433594,123.64581298828125],"size":[349.99859619140625,106],"flags":{},"order":4,"mode":0,"inputs":[],"outputs":[{"name":"GEMMA","type":"GEMMA","links":[8,10],"slot_index":0}],"properties":{"Node name for S&R":"GemmaLoader"},"widgets_values":["unsloth/gemma-2-2b-it-bnb-4bit","cpu","default"]}],"links":[[1,3,0,6,0,"MODEL"],[2,4,0,2,1,"INT"],[3,4,1,2,2,"INT"],[4,4,0,5,0,"INT"],[5,4,1,5,1,"INT"],[8,8,0,1,0,"GEMMA"],[10,8,0,9,0,"GEMMA"],[11,9,0,6,2,"CONDITIONING"],[12,1,0,2,0,"CONDITIONING"],[13,2,0,6,1,"CONDITIONING"],[14,6,0,10,0,"LATENT"],[15,11,0,10,1,"VAE"],[16,10,0,12,0,"IMAGE"],[20,5,0,6,3,"LATENT"],[21,6,0,15,3,"LATENT"],[22,15,0,16,0,"LATENT"],[23,11,0,16,1,"VAE"],[28,3,0,15,0,"MODEL"],[29,16,0,19,0,"IMAGE"],[30,20,0,15,4,"INT"],[31,20,0,6,4,"INT"],[32,9,0,15,2,"CONDITIONING"],[33,1,0,15,1,"CONDITIONING"]],"groups":[{"id":1,"title":"Group","bounding":[192.7167205810547,1017.5338134765625,668.3709716796875,566.7073364257812],"color":"#3f789e","font_size":24,"flags":{}}],"config":{},"extra":{"ds":{"scale":0.8438022155513367,"offset":[668.366620908559,115.06668170315389]},"ue_links":[]},"version":0.4}

Additional Context

(Please add any additional context or steps to reproduce the error here)

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions