-
Notifications
You must be signed in to change notification settings - Fork 59
Open
Description
GemmaLoader
The checkpoint you are trying to load has model type gemma2 but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
ComfyUI Error Report
Error Details
- Node ID: 8
- Node Type: GemmaLoader
- Exception Type: ValueError
- Exception Message: The checkpoint you are trying to load has model type
gemma2but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
Stack Trace
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
raise ValueError(
System Information
- ComfyUI Version: v0.3.5-2-g497db62
- Arguments: ComfyUI\main.py --windows-standalone-build
- OS: nt
- Python Version: 3.12.7 (tags/v3.12.7:0b05ead, Oct 1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]
- Embedded Python: true
- PyTorch Version: 2.5.1+cu124
Devices
- Name: cuda:0 NVIDIA GeForce RTX 3060 : cudaMallocAsync
- Type: cuda
- VRAM Total: 12884246528
- VRAM Free: 11788091392
- Torch VRAM Total: 0
- Torch VRAM Free: 0
Logs
2025-01-03T00:02:05.248361 - [START] Security scan2025-01-03T00:02:05.248361 -
2025-01-03T00:02:11.680609 - [DONE] Security scan2025-01-03T00:02:11.680609 -
2025-01-03T00:03:05.240208 - Starting server
2025-01-03T00:03:05.240208 - To see the GUI go to: http://127.0.0.1:8188
2025-01-03T00:03:11.924211 - FETCH DATA from: D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json2025-01-03T00:03:11.951207 - [DONE]2025-01-03T00:03:11.952207 -
2025-01-03T00:03:24.934161 - got prompt
2025-01-03T00:03:28.107182 - Missing VAE keys2025-01-03T00:03:28.108184 - 2025-01-03T00:03:28.108184 - ['encoder.project_in.weight', 'encoder.project_in.bias', 'encoder.stages.0.0.conv1.conv.weight', 'encoder.stages.0.0.conv1.conv.bias', 'encoder.stages.0.0.conv2.conv.weight', 'encoder.stages.0.0.conv2.norm.weight', 'encoder.stages.0.0.conv2.norm.bias', 'encoder.stages.0.1.conv1.conv.weight', 'encoder.stages.0.1.conv1.conv.bias', 'encoder.stages.0.1.conv2.conv.weight', 'encoder.stages.0.1.conv2.norm.weight', 'encoder.stages.0.1.conv2.norm.bias', 'encoder.stages.0.2.main.weight', 'encoder.stages.0.2.main.bias', 'encoder.stages.1.0.conv1.conv.weight', 'encoder.stages.1.0.conv1.conv.bias', 'encoder.stages.1.0.conv2.conv.weight', 'encoder.stages.1.0.conv2.norm.weight', 'encoder.stages.1.0.conv2.norm.bias', 'encoder.stages.1.1.conv1.conv.weight', 'encoder.stages.1.1.conv1.conv.bias', 'encoder.stages.1.1.conv2.conv.weight', 'encoder.stages.1.1.conv2.norm.weight', 'encoder.stages.1.1.conv2.norm.bias', 'encoder.stages.1.2.main.weight', 'encoder.stages.1.2.main.bias', 'encoder.stages.2.0.conv1.conv.weight', 'encoder.stages.2.0.conv1.conv.bias', 'encoder.stages.2.0.conv2.conv.weight', 'encoder.stages.2.0.conv2.norm.weight', 'encoder.stages.2.0.conv2.norm.bias', 'encoder.stages.2.1.conv1.conv.weight', 'encoder.stages.2.1.conv1.conv.bias', 'encoder.stages.2.1.conv2.conv.weight', 'encoder.stages.2.1.conv2.norm.weight', 'encoder.stages.2.1.conv2.norm.bias', 'encoder.stages.2.2.main.weight', 'encoder.stages.2.2.main.bias', 'encoder.stages.3.0.context_module.qkv.0.weight', 'encoder.stages.3.0.context_module.aggreg.0.0.weight', 'encoder.stages.3.0.context_module.aggreg.0.1.weight', 'encoder.stages.3.0.context_module.proj.0.weight', 'encoder.stages.3.0.context_module.proj.1.weight', 'encoder.stages.3.0.context_module.proj.1.bias', 'encoder.stages.3.0.local_module.inverted_conv.conv.weight', 'encoder.stages.3.0.local_module.inverted_conv.conv.bias', 'encoder.stages.3.0.local_module.depth_conv.conv.weight', 'encoder.stages.3.0.local_module.depth_conv.conv.bias', 'encoder.stages.3.0.local_module.point_conv.conv.weight', 'encoder.stages.3.0.local_module.point_conv.norm.weight', 'encoder.stages.3.0.local_module.point_conv.norm.bias', 'encoder.stages.3.1.context_module.qkv.0.weight', 'encoder.stages.3.1.context_module.aggreg.0.0.weight', 'encoder.stages.3.1.context_module.aggreg.0.1.weight', 'encoder.stages.3.1.context_module.proj.0.weight', 'encoder.stages.3.1.context_module.proj.1.weight', 'encoder.stages.3.1.context_module.proj.1.bias', 'encoder.stages.3.1.local_module.inverted_conv.conv.weight', 'encoder.stages.3.1.local_module.inverted_conv.conv.bias', 'encoder.stages.3.1.local_module.depth_conv.conv.weight', 'encoder.stages.3.1.local_module.depth_conv.conv.bias', 'encoder.stages.3.1.local_module.point_conv.conv.weight', 'encoder.stages.3.1.local_module.point_conv.norm.weight', 'encoder.stages.3.1.local_module.point_conv.norm.bias', 'encoder.stages.3.2.context_module.qkv.0.weight', 'encoder.stages.3.2.context_module.aggreg.0.0.weight', 'encoder.stages.3.2.context_module.aggreg.0.1.weight', 'encoder.stages.3.2.context_module.proj.0.weight', 'encoder.stages.3.2.context_module.proj.1.weight', 'encoder.stages.3.2.context_module.proj.1.bias', 'encoder.stages.3.2.local_module.inverted_conv.conv.weight', 'encoder.stages.3.2.local_module.inverted_conv.conv.bias', 'encoder.stages.3.2.local_module.depth_conv.conv.weight', 'encoder.stages.3.2.local_module.depth_conv.conv.bias', 'encoder.stages.3.2.local_module.point_conv.conv.weight', 'encoder.stages.3.2.local_module.point_conv.norm.weight', 'encoder.stages.3.2.local_module.point_conv.norm.bias', 'encoder.stages.3.3.main.weight', 'encoder.stages.3.3.main.bias', 'encoder.stages.4.0.context_module.qkv.0.weight', 'encoder.stages.4.0.context_module.aggreg.0.0.weight', 'encoder.stages.4.0.context_module.aggreg.0.1.weight', 'encoder.stages.4.0.context_module.proj.0.weight', 'encoder.stages.4.0.context_module.proj.1.weight', 'encoder.stages.4.0.context_module.proj.1.bias', 'encoder.stages.4.0.local_module.inverted_conv.conv.weight', 'encoder.stages.4.0.local_module.inverted_conv.conv.bias', 'encoder.stages.4.0.local_module.depth_conv.conv.weight', 'encoder.stages.4.0.local_module.depth_conv.conv.bias', 'encoder.stages.4.0.local_module.point_conv.conv.weight', 'encoder.stages.4.0.local_module.point_conv.norm.weight', 'encoder.stages.4.0.local_module.point_conv.norm.bias', 'encoder.stages.4.1.context_module.qkv.0.weight', 'encoder.stages.4.1.context_module.aggreg.0.0.weight', 'encoder.stages.4.1.context_module.aggreg.0.1.weight', 'encoder.stages.4.1.context_module.proj.0.weight', 'encoder.stages.4.1.context_module.proj.1.weight', 'encoder.stages.4.1.context_module.proj.1.bias', 'encoder.stages.4.1.local_module.inverted_conv.conv.weight', 'encoder.stages.4.1.local_module.inverted_conv.conv.bias', 'encoder.stages.4.1.local_module.depth_conv.conv.weight', 'encoder.stages.4.1.local_module.depth_conv.conv.bias', 'encoder.stages.4.1.local_module.point_conv.conv.weight', 'encoder.stages.4.1.local_module.point_conv.norm.weight', 'encoder.stages.4.1.local_module.point_conv.norm.bias', 'encoder.stages.4.2.context_module.qkv.0.weight', 'encoder.stages.4.2.context_module.aggreg.0.0.weight', 'encoder.stages.4.2.context_module.aggreg.0.1.weight', 'encoder.stages.4.2.context_module.proj.0.weight', 'encoder.stages.4.2.context_module.proj.1.weight', 'encoder.stages.4.2.context_module.proj.1.bias', 'encoder.stages.4.2.local_module.inverted_conv.conv.weight', 'encoder.stages.4.2.local_module.inverted_conv.conv.bias', 'encoder.stages.4.2.local_module.depth_conv.conv.weight', 'encoder.stages.4.2.local_module.depth_conv.conv.bias', 'encoder.stages.4.2.local_module.point_conv.conv.weight', 'encoder.stages.4.2.local_module.point_conv.norm.weight', 'encoder.stages.4.2.local_module.point_conv.norm.bias', 'encoder.stages.4.3.main.weight', 'encoder.stages.4.3.main.bias', 'encoder.stages.5.0.context_module.qkv.0.weight', 'encoder.stages.5.0.context_module.aggreg.0.0.weight', 'encoder.stages.5.0.context_module.aggreg.0.1.weight', 'encoder.stages.5.0.context_module.proj.0.weight', 'encoder.stages.5.0.context_module.proj.1.weight', 'encoder.stages.5.0.context_module.proj.1.bias', 'encoder.stages.5.0.local_module.inverted_conv.conv.weight', 'encoder.stages.5.0.local_module.inverted_conv.conv.bias', 'encoder.stages.5.0.local_module.depth_conv.conv.weight', 'encoder.stages.5.0.local_module.depth_conv.conv.bias', 'encoder.stages.5.0.local_module.point_conv.conv.weight', 'encoder.stages.5.0.local_module.point_conv.norm.weight', 'encoder.stages.5.0.local_module.point_conv.norm.bias', 'encoder.stages.5.1.context_module.qkv.0.weight', 'encoder.stages.5.1.context_module.aggreg.0.0.weight', 'encoder.stages.5.1.context_module.aggreg.0.1.weight', 'encoder.stages.5.1.context_module.proj.0.weight', 'encoder.stages.5.1.context_module.proj.1.weight', 'encoder.stages.5.1.context_module.proj.1.bias', 'encoder.stages.5.1.local_module.inverted_conv.conv.weight', 'encoder.stages.5.1.local_module.inverted_conv.conv.bias', 'encoder.stages.5.1.local_module.depth_conv.conv.weight', 'encoder.stages.5.1.local_module.depth_conv.conv.bias', 'encoder.stages.5.1.local_module.point_conv.conv.weight', 'encoder.stages.5.1.local_module.point_conv.norm.weight', 'encoder.stages.5.1.local_module.point_conv.norm.bias', 'encoder.stages.5.2.context_module.qkv.0.weight', 'encoder.stages.5.2.context_module.aggreg.0.0.weight', 'encoder.stages.5.2.context_module.aggreg.0.1.weight', 'encoder.stages.5.2.context_module.proj.0.weight', 'encoder.stages.5.2.context_module.proj.1.weight', 'encoder.stages.5.2.context_module.proj.1.bias', 'encoder.stages.5.2.local_module.inverted_conv.conv.weight', 'encoder.stages.5.2.local_module.inverted_conv.conv.bias', 'encoder.stages.5.2.local_module.depth_conv.conv.weight', 'encoder.stages.5.2.local_module.depth_conv.conv.bias', 'encoder.stages.5.2.local_module.point_conv.conv.weight', 'encoder.stages.5.2.local_module.point_conv.norm.weight', 'encoder.stages.5.2.local_module.point_conv.norm.bias', 'encoder.project_out.main.0.conv.weight', 'encoder.project_out.main.0.conv.bias', 'decoder.project_in.main.conv.weight', 'decoder.project_in.main.conv.bias', 'decoder.stages.0.0.main.conv.weight', 'decoder.stages.0.0.main.conv.bias', 'decoder.stages.0.1.conv1.conv.weight', 'decoder.stages.0.1.conv1.conv.bias', 'decoder.stages.0.1.conv2.conv.weight', 'decoder.stages.0.1.conv2.norm.weight', 'decoder.stages.0.1.conv2.norm.bias', 'decoder.stages.0.2.conv1.conv.weight', 'decoder.stages.0.2.conv1.conv.bias', 'decoder.stages.0.2.conv2.conv.weight', 'decoder.stages.0.2.conv2.norm.weight', 'decoder.stages.0.2.conv2.norm.bias', 'decoder.stages.0.3.conv1.conv.weight', 'decoder.stages.0.3.conv1.conv.bias', 'decoder.stages.0.3.conv2.conv.weight', 'decoder.stages.0.3.conv2.norm.weight', 'decoder.stages.0.3.conv2.norm.bias', 'decoder.stages.1.0.main.conv.weight', 'decoder.stages.1.0.main.conv.bias', 'decoder.stages.1.1.conv1.conv.weight', 'decoder.stages.1.1.conv1.conv.bias', 'decoder.stages.1.1.conv2.conv.weight', 'decoder.stages.1.1.conv2.norm.weight', 'decoder.stages.1.1.conv2.norm.bias', 'decoder.stages.1.2.conv1.conv.weight', 'decoder.stages.1.2.conv1.conv.bias', 'decoder.stages.1.2.conv2.conv.weight', 'decoder.stages.1.2.conv2.norm.weight', 'decoder.stages.1.2.conv2.norm.bias', 'decoder.stages.1.3.conv1.conv.weight', 'decoder.stages.1.3.conv1.conv.bias', 'decoder.stages.1.3.conv2.conv.weight', 'decoder.stages.1.3.conv2.norm.weight', 'decoder.stages.1.3.conv2.norm.bias', 'decoder.stages.2.0.main.conv.weight', 'decoder.stages.2.0.main.conv.bias', 'decoder.stages.2.1.conv1.conv.weight', 'decoder.stages.2.1.conv1.conv.bias', 'decoder.stages.2.1.conv2.conv.weight', 'decoder.stages.2.1.conv2.norm.weight', 'decoder.stages.2.1.conv2.norm.bias', 'decoder.stages.2.2.conv1.conv.weight', 'decoder.stages.2.2.conv1.conv.bias', 'decoder.stages.2.2.conv2.conv.weight', 'decoder.stages.2.2.conv2.norm.weight', 'decoder.stages.2.2.conv2.norm.bias', 'decoder.stages.2.3.conv1.conv.weight', 'decoder.stages.2.3.conv1.conv.bias', 'decoder.stages.2.3.conv2.conv.weight', 'decoder.stages.2.3.conv2.norm.weight', 'decoder.stages.2.3.conv2.norm.bias', 'decoder.stages.3.0.main.conv.weight', 'decoder.stages.3.0.main.conv.bias', 'decoder.stages.3.1.context_module.qkv.0.weight', 'decoder.stages.3.1.context_module.aggreg.0.0.weight', 'decoder.stages.3.1.context_module.aggreg.0.1.weight', 'decoder.stages.3.1.context_module.proj.0.weight', 'decoder.stages.3.1.context_module.proj.1.weight', 'decoder.stages.3.1.context_module.proj.1.bias', 'decoder.stages.3.1.local_module.inverted_conv.conv.weight', 'decoder.stages.3.1.local_module.inverted_conv.conv.bias', 'decoder.stages.3.1.local_module.depth_conv.conv.weight', 'decoder.stages.3.1.local_module.depth_conv.conv.bias', 'decoder.stages.3.1.local_module.point_conv.conv.weight', 'decoder.stages.3.1.local_module.point_conv.norm.weight', 'decoder.stages.3.1.local_module.point_conv.norm.bias', 'decoder.stages.3.2.context_module.qkv.0.weight', 'decoder.stages.3.2.context_module.aggreg.0.0.weight', 'decoder.stages.3.2.context_module.aggreg.0.1.weight', 'decoder.stages.3.2.context_module.proj.0.weight', 'decoder.stages.3.2.context_module.proj.1.weight', 'decoder.stages.3.2.context_module.proj.1.bias', 'decoder.stages.3.2.local_module.inverted_conv.conv.weight', 'decoder.stages.3.2.local_module.inverted_conv.conv.bias', 'decoder.stages.3.2.local_module.depth_conv.conv.weight', 'decoder.stages.3.2.local_module.depth_conv.conv.bias', 'decoder.stages.3.2.local_module.point_conv.conv.weight', 'decoder.stages.3.2.local_module.point_conv.norm.weight', 'decoder.stages.3.2.local_module.point_conv.norm.bias', 'decoder.stages.3.3.context_module.qkv.0.weight', 'decoder.stages.3.3.context_module.aggreg.0.0.weight', 'decoder.stages.3.3.context_module.aggreg.0.1.weight', 'decoder.stages.3.3.context_module.proj.0.weight', 'decoder.stages.3.3.context_module.proj.1.weight', 'decoder.stages.3.3.context_module.proj.1.bias', 'decoder.stages.3.3.local_module.inverted_conv.conv.weight', 'decoder.stages.3.3.local_module.inverted_conv.conv.bias', 'decoder.stages.3.3.local_module.depth_conv.conv.weight', 'decoder.stages.3.3.local_module.depth_conv.conv.bias', 'decoder.stages.3.3.local_module.point_conv.conv.weight', 'decoder.stages.3.3.local_module.point_conv.norm.weight', 'decoder.stages.3.3.local_module.point_conv.norm.bias', 'decoder.stages.4.0.main.conv.weight', 'decoder.stages.4.0.main.conv.bias', 'decoder.stages.4.1.context_module.qkv.0.weight', 'decoder.stages.4.1.context_module.aggreg.0.0.weight', 'decoder.stages.4.1.context_module.aggreg.0.1.weight', 'decoder.stages.4.1.context_module.proj.0.weight', 'decoder.stages.4.1.context_module.proj.1.weight', 'decoder.stages.4.1.context_module.proj.1.bias', 'decoder.stages.4.1.local_module.inverted_conv.conv.weight', 'decoder.stages.4.1.local_module.inverted_conv.conv.bias', 'decoder.stages.4.1.local_module.depth_conv.conv.weight', 'decoder.stages.4.1.local_module.depth_conv.conv.bias', 'decoder.stages.4.1.local_module.point_conv.conv.weight', 'decoder.stages.4.1.local_module.point_conv.norm.weight', 'decoder.stages.4.1.local_module.point_conv.norm.bias', 'decoder.stages.4.2.context_module.qkv.0.weight', 'decoder.stages.4.2.context_module.aggreg.0.0.weight', 'decoder.stages.4.2.context_module.aggreg.0.1.weight', 'decoder.stages.4.2.context_module.proj.0.weight', 'decoder.stages.4.2.context_module.proj.1.weight', 'decoder.stages.4.2.context_module.proj.1.bias', 'decoder.stages.4.2.local_module.inverted_conv.conv.weight', 'decoder.stages.4.2.local_module.inverted_conv.conv.bias', 'decoder.stages.4.2.local_module.depth_conv.conv.weight', 'decoder.stages.4.2.local_module.depth_conv.conv.bias', 'decoder.stages.4.2.local_module.point_conv.conv.weight', 'decoder.stages.4.2.local_module.point_conv.norm.weight', 'decoder.stages.4.2.local_module.point_conv.norm.bias', 'decoder.stages.4.3.context_module.qkv.0.weight', 'decoder.stages.4.3.context_module.aggreg.0.0.weight', 'decoder.stages.4.3.context_module.aggreg.0.1.weight', 'decoder.stages.4.3.context_module.proj.0.weight', 'decoder.stages.4.3.context_module.proj.1.weight', 'decoder.stages.4.3.context_module.proj.1.bias', 'decoder.stages.4.3.local_module.inverted_conv.conv.weight', 'decoder.stages.4.3.local_module.inverted_conv.conv.bias', 'decoder.stages.4.3.local_module.depth_conv.conv.weight', 'decoder.stages.4.3.local_module.depth_conv.conv.bias', 'decoder.stages.4.3.local_module.point_conv.conv.weight', 'decoder.stages.4.3.local_module.point_conv.norm.weight', 'decoder.stages.4.3.local_module.point_conv.norm.bias', 'decoder.stages.5.0.context_module.qkv.0.weight', 'decoder.stages.5.0.context_module.aggreg.0.0.weight', 'decoder.stages.5.0.context_module.aggreg.0.1.weight', 'decoder.stages.5.0.context_module.proj.0.weight', 'decoder.stages.5.0.context_module.proj.1.weight', 'decoder.stages.5.0.context_module.proj.1.bias', 'decoder.stages.5.0.local_module.inverted_conv.conv.weight', 'decoder.stages.5.0.local_module.inverted_conv.conv.bias', 'decoder.stages.5.0.local_module.depth_conv.conv.weight', 'decoder.stages.5.0.local_module.depth_conv.conv.bias', 'decoder.stages.5.0.local_module.point_conv.conv.weight', 'decoder.stages.5.0.local_module.point_conv.norm.weight', 'decoder.stages.5.0.local_module.point_conv.norm.bias', 'decoder.stages.5.1.context_module.qkv.0.weight', 'decoder.stages.5.1.context_module.aggreg.0.0.weight', 'decoder.stages.5.1.context_module.aggreg.0.1.weight', 'decoder.stages.5.1.context_module.proj.0.weight', 'decoder.stages.5.1.context_module.proj.1.weight', 'decoder.stages.5.1.context_module.proj.1.bias', 'decoder.stages.5.1.local_module.inverted_conv.conv.weight', 'decoder.stages.5.1.local_module.inverted_conv.conv.bias', 'decoder.stages.5.1.local_module.depth_conv.conv.weight', 'decoder.stages.5.1.local_module.depth_conv.conv.bias', 'decoder.stages.5.1.local_module.point_conv.conv.weight', 'decoder.stages.5.1.local_module.point_conv.norm.weight', 'decoder.stages.5.1.local_module.point_conv.norm.bias', 'decoder.stages.5.2.context_module.qkv.0.weight', 'decoder.stages.5.2.context_module.aggreg.0.0.weight', 'decoder.stages.5.2.context_module.aggreg.0.1.weight', 'decoder.stages.5.2.context_module.proj.0.weight', 'decoder.stages.5.2.context_module.proj.1.weight', 'decoder.stages.5.2.context_module.proj.1.bias', 'decoder.stages.5.2.local_module.inverted_conv.conv.weight', 'decoder.stages.5.2.local_module.inverted_conv.conv.bias', 'decoder.stages.5.2.local_module.depth_conv.conv.weight', 'decoder.stages.5.2.local_module.depth_conv.conv.bias', 'decoder.stages.5.2.local_module.point_conv.conv.weight', 'decoder.stages.5.2.local_module.point_conv.norm.weight', 'decoder.stages.5.2.local_module.point_conv.norm.bias', 'decoder.project_out.0.weight', 'decoder.project_out.0.bias', 'decoder.project_out.2.conv.weight', 'decoder.project_out.2.conv.bias']2025-01-03T00:03:28.111183 -
2025-01-03T00:03:28.111183 - Leftover VAE keys2025-01-03T00:03:28.111183 - 2025-01-03T00:03:28.111183 - ['encoder.conv_in.bias', 'encoder.conv_in.weight', 'encoder.conv_out.bias', 'encoder.conv_out.weight', 'encoder.down_blocks.0.0.conv1.bias', 'encoder.down_blocks.0.0.conv1.weight', 'encoder.down_blocks.0.0.conv2.weight', 'encoder.down_blocks.0.0.norm.bias', 'encoder.down_blocks.0.0.norm.weight', 'encoder.down_blocks.0.1.conv1.bias', 'encoder.down_blocks.0.1.conv1.weight', 'encoder.down_blocks.0.1.conv2.weight', 'encoder.down_blocks.0.1.norm.bias', 'encoder.down_blocks.0.1.norm.weight', 'encoder.down_blocks.0.2.conv.bias', 'encoder.down_blocks.0.2.conv.weight', 'encoder.down_blocks.1.0.conv1.bias', 'encoder.down_blocks.1.0.conv1.weight', 'encoder.down_blocks.1.0.conv2.weight', 'encoder.down_blocks.1.0.norm.bias', 'encoder.down_blocks.1.0.norm.weight', 'encoder.down_blocks.1.1.conv1.bias', 'encoder.down_blocks.1.1.conv1.weight', 'encoder.down_blocks.1.1.conv2.weight', 'encoder.down_blocks.1.1.norm.bias', 'encoder.down_blocks.1.1.norm.weight', 'encoder.down_blocks.1.2.conv.bias', 'encoder.down_blocks.1.2.conv.weight', 'encoder.down_blocks.2.0.conv1.bias', 'encoder.down_blocks.2.0.conv1.weight', 'encoder.down_blocks.2.0.conv2.weight', 'encoder.down_blocks.2.0.norm.bias', 'encoder.down_blocks.2.0.norm.weight', 'encoder.down_blocks.2.1.conv1.bias', 'encoder.down_blocks.2.1.conv1.weight', 'encoder.down_blocks.2.1.conv2.weight', 'encoder.down_blocks.2.1.norm.bias', 'encoder.down_blocks.2.1.norm.weight', 'encoder.down_blocks.2.2.conv.bias', 'encoder.down_blocks.2.2.conv.weight', 'encoder.down_blocks.3.0.attn.norm_out.bias', 'encoder.down_blocks.3.0.attn.norm_out.weight', 'encoder.down_blocks.3.0.attn.to_k.weight', 'encoder.down_blocks.3.0.attn.to_out.weight', 'encoder.down_blocks.3.0.attn.to_q.weight', 'encoder.down_blocks.3.0.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.3.0.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.3.0.attn.to_v.weight', 'encoder.down_blocks.3.0.conv_out.conv_depth.bias', 'encoder.down_blocks.3.0.conv_out.conv_depth.weight', 'encoder.down_blocks.3.0.conv_out.conv_inverted.bias', 'encoder.down_blocks.3.0.conv_out.conv_inverted.weight', 'encoder.down_blocks.3.0.conv_out.conv_point.weight', 'encoder.down_blocks.3.0.conv_out.norm.bias', 'encoder.down_blocks.3.0.conv_out.norm.weight', 'encoder.down_blocks.3.1.attn.norm_out.bias', 'encoder.down_blocks.3.1.attn.norm_out.weight', 'encoder.down_blocks.3.1.attn.to_k.weight', 'encoder.down_blocks.3.1.attn.to_out.weight', 'encoder.down_blocks.3.1.attn.to_q.weight', 'encoder.down_blocks.3.1.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.3.1.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.3.1.attn.to_v.weight', 'encoder.down_blocks.3.1.conv_out.conv_depth.bias', 'encoder.down_blocks.3.1.conv_out.conv_depth.weight', 'encoder.down_blocks.3.1.conv_out.conv_inverted.bias', 'encoder.down_blocks.3.1.conv_out.conv_inverted.weight', 'encoder.down_blocks.3.1.conv_out.conv_point.weight', 'encoder.down_blocks.3.1.conv_out.norm.bias', 'encoder.down_blocks.3.1.conv_out.norm.weight', 'encoder.down_blocks.3.2.attn.norm_out.bias', 'encoder.down_blocks.3.2.attn.norm_out.weight', 'encoder.down_blocks.3.2.attn.to_k.weight', 'encoder.down_blocks.3.2.attn.to_out.weight', 'encoder.down_blocks.3.2.attn.to_q.weight', 'encoder.down_blocks.3.2.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.3.2.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.3.2.attn.to_v.weight', 'encoder.down_blocks.3.2.conv_out.conv_depth.bias', 'encoder.down_blocks.3.2.conv_out.conv_depth.weight', 'encoder.down_blocks.3.2.conv_out.conv_inverted.bias', 'encoder.down_blocks.3.2.conv_out.conv_inverted.weight', 'encoder.down_blocks.3.2.conv_out.conv_point.weight', 'encoder.down_blocks.3.2.conv_out.norm.bias', 'encoder.down_blocks.3.2.conv_out.norm.weight', 'encoder.down_blocks.3.3.conv.bias', 'encoder.down_blocks.3.3.conv.weight', 'encoder.down_blocks.4.0.attn.norm_out.bias', 'encoder.down_blocks.4.0.attn.norm_out.weight', 'encoder.down_blocks.4.0.attn.to_k.weight', 'encoder.down_blocks.4.0.attn.to_out.weight', 'encoder.down_blocks.4.0.attn.to_q.weight', 'encoder.down_blocks.4.0.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.4.0.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.4.0.attn.to_v.weight', 'encoder.down_blocks.4.0.conv_out.conv_depth.bias', 'encoder.down_blocks.4.0.conv_out.conv_depth.weight', 'encoder.down_blocks.4.0.conv_out.conv_inverted.bias', 'encoder.down_blocks.4.0.conv_out.conv_inverted.weight', 'encoder.down_blocks.4.0.conv_out.conv_point.weight', 'encoder.down_blocks.4.0.conv_out.norm.bias', 'encoder.down_blocks.4.0.conv_out.norm.weight', 'encoder.down_blocks.4.1.attn.norm_out.bias', 'encoder.down_blocks.4.1.attn.norm_out.weight', 'encoder.down_blocks.4.1.attn.to_k.weight', 'encoder.down_blocks.4.1.attn.to_out.weight', 'encoder.down_blocks.4.1.attn.to_q.weight', 'encoder.down_blocks.4.1.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.4.1.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.4.1.attn.to_v.weight', 'encoder.down_blocks.4.1.conv_out.conv_depth.bias', 'encoder.down_blocks.4.1.conv_out.conv_depth.weight', 'encoder.down_blocks.4.1.conv_out.conv_inverted.bias', 'encoder.down_blocks.4.1.conv_out.conv_inverted.weight', 'encoder.down_blocks.4.1.conv_out.conv_point.weight', 'encoder.down_blocks.4.1.conv_out.norm.bias', 'encoder.down_blocks.4.1.conv_out.norm.weight', 'encoder.down_blocks.4.2.attn.norm_out.bias', 'encoder.down_blocks.4.2.attn.norm_out.weight', 'encoder.down_blocks.4.2.attn.to_k.weight', 'encoder.down_blocks.4.2.attn.to_out.weight', 'encoder.down_blocks.4.2.attn.to_q.weight', 'encoder.down_blocks.4.2.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.4.2.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.4.2.attn.to_v.weight', 'encoder.down_blocks.4.2.conv_out.conv_depth.bias', 'encoder.down_blocks.4.2.conv_out.conv_depth.weight', 'encoder.down_blocks.4.2.conv_out.conv_inverted.bias', 'encoder.down_blocks.4.2.conv_out.conv_inverted.weight', 'encoder.down_blocks.4.2.conv_out.conv_point.weight', 'encoder.down_blocks.4.2.conv_out.norm.bias', 'encoder.down_blocks.4.2.conv_out.norm.weight', 'encoder.down_blocks.4.3.conv.bias', 'encoder.down_blocks.4.3.conv.weight', 'encoder.down_blocks.5.0.attn.norm_out.bias', 'encoder.down_blocks.5.0.attn.norm_out.weight', 'encoder.down_blocks.5.0.attn.to_k.weight', 'encoder.down_blocks.5.0.attn.to_out.weight', 'encoder.down_blocks.5.0.attn.to_q.weight', 'encoder.down_blocks.5.0.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.5.0.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.5.0.attn.to_v.weight', 'encoder.down_blocks.5.0.conv_out.conv_depth.bias', 'encoder.down_blocks.5.0.conv_out.conv_depth.weight', 'encoder.down_blocks.5.0.conv_out.conv_inverted.bias', 'encoder.down_blocks.5.0.conv_out.conv_inverted.weight', 'encoder.down_blocks.5.0.conv_out.conv_point.weight', 'encoder.down_blocks.5.0.conv_out.norm.bias', 'encoder.down_blocks.5.0.conv_out.norm.weight', 'encoder.down_blocks.5.1.attn.norm_out.bias', 'encoder.down_blocks.5.1.attn.norm_out.weight', 'encoder.down_blocks.5.1.attn.to_k.weight', 'encoder.down_blocks.5.1.attn.to_out.weight', 'encoder.down_blocks.5.1.attn.to_q.weight', 'encoder.down_blocks.5.1.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.5.1.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.5.1.attn.to_v.weight', 'encoder.down_blocks.5.1.conv_out.conv_depth.bias', 'encoder.down_blocks.5.1.conv_out.conv_depth.weight', 'encoder.down_blocks.5.1.conv_out.conv_inverted.bias', 'encoder.down_blocks.5.1.conv_out.conv_inverted.weight', 'encoder.down_blocks.5.1.conv_out.conv_point.weight', 'encoder.down_blocks.5.1.conv_out.norm.bias', 'encoder.down_blocks.5.1.conv_out.norm.weight', 'encoder.down_blocks.5.2.attn.norm_out.bias', 'encoder.down_blocks.5.2.attn.norm_out.weight', 'encoder.down_blocks.5.2.attn.to_k.weight', 'encoder.down_blocks.5.2.attn.to_out.weight', 'encoder.down_blocks.5.2.attn.to_q.weight', 'encoder.down_blocks.5.2.attn.to_qkv_multiscale.0.proj_in.weight', 'encoder.down_blocks.5.2.attn.to_qkv_multiscale.0.proj_out.weight', 'encoder.down_blocks.5.2.attn.to_v.weight', 'encoder.down_blocks.5.2.conv_out.conv_depth.bias', 'encoder.down_blocks.5.2.conv_out.conv_depth.weight', 'encoder.down_blocks.5.2.conv_out.conv_inverted.bias', 'encoder.down_blocks.5.2.conv_out.conv_inverted.weight', 'encoder.down_blocks.5.2.conv_out.conv_point.weight', 'encoder.down_blocks.5.2.conv_out.norm.bias', 'encoder.down_blocks.5.2.conv_out.norm.weight', 'decoder.conv_in.bias', 'decoder.conv_in.weight', 'decoder.conv_out.bias', 'decoder.conv_out.weight', 'decoder.norm_out.bias', 'decoder.norm_out.weight', 'decoder.up_blocks.0.0.conv.bias', 'decoder.up_blocks.0.0.conv.weight', 'decoder.up_blocks.0.1.conv1.bias', 'decoder.up_blocks.0.1.conv1.weight', 'decoder.up_blocks.0.1.conv2.weight', 'decoder.up_blocks.0.1.norm.bias', 'decoder.up_blocks.0.1.norm.weight', 'decoder.up_blocks.0.2.conv1.bias', 'decoder.up_blocks.0.2.conv1.weight', 'decoder.up_blocks.0.2.conv2.weight', 'decoder.up_blocks.0.2.norm.bias', 'decoder.up_blocks.0.2.norm.weight', 'decoder.up_blocks.0.3.conv1.bias', 'decoder.up_blocks.0.3.conv1.weight', 'decoder.up_blocks.0.3.conv2.weight', 'decoder.up_blocks.0.3.norm.bias', 'decoder.up_blocks.0.3.norm.weight', 'decoder.up_blocks.1.0.conv.bias', 'decoder.up_blocks.1.0.conv.weight', 'decoder.up_blocks.1.1.conv1.bias', 'decoder.up_blocks.1.1.conv1.weight', 'decoder.up_blocks.1.1.conv2.weight', 'decoder.up_blocks.1.1.norm.bias', 'decoder.up_blocks.1.1.norm.weight', 'decoder.up_blocks.1.2.conv1.bias', 'decoder.up_blocks.1.2.conv1.weight', 'decoder.up_blocks.1.2.conv2.weight', 'decoder.up_blocks.1.2.norm.bias', 'decoder.up_blocks.1.2.norm.weight', 'decoder.up_blocks.1.3.conv1.bias', 'decoder.up_blocks.1.3.conv1.weight', 'decoder.up_blocks.1.3.conv2.weight', 'decoder.up_blocks.1.3.norm.bias', 'decoder.up_blocks.1.3.norm.weight', 'decoder.up_blocks.2.0.conv.bias', 'decoder.up_blocks.2.0.conv.weight', 'decoder.up_blocks.2.1.conv1.bias', 'decoder.up_blocks.2.1.conv1.weight', 'decoder.up_blocks.2.1.conv2.weight', 'decoder.up_blocks.2.1.norm.bias', 'decoder.up_blocks.2.1.norm.weight', 'decoder.up_blocks.2.2.conv1.bias', 'decoder.up_blocks.2.2.conv1.weight', 'decoder.up_blocks.2.2.conv2.weight', 'decoder.up_blocks.2.2.norm.bias', 'decoder.up_blocks.2.2.norm.weight', 'decoder.up_blocks.2.3.conv1.bias', 'decoder.up_blocks.2.3.conv1.weight', 'decoder.up_blocks.2.3.conv2.weight', 'decoder.up_blocks.2.3.norm.bias', 'decoder.up_blocks.2.3.norm.weight', 'decoder.up_blocks.3.0.conv.bias', 'decoder.up_blocks.3.0.conv.weight', 'decoder.up_blocks.3.1.attn.norm_out.bias', 'decoder.up_blocks.3.1.attn.norm_out.weight', 'decoder.up_blocks.3.1.attn.to_k.weight', 'decoder.up_blocks.3.1.attn.to_out.weight', 'decoder.up_blocks.3.1.attn.to_q.weight', 'decoder.up_blocks.3.1.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.3.1.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.3.1.attn.to_v.weight', 'decoder.up_blocks.3.1.conv_out.conv_depth.bias', 'decoder.up_blocks.3.1.conv_out.conv_depth.weight', 'decoder.up_blocks.3.1.conv_out.conv_inverted.bias', 'decoder.up_blocks.3.1.conv_out.conv_inverted.weight', 'decoder.up_blocks.3.1.conv_out.conv_point.weight', 'decoder.up_blocks.3.1.conv_out.norm.bias', 'decoder.up_blocks.3.1.conv_out.norm.weight', 'decoder.up_blocks.3.2.attn.norm_out.bias', 'decoder.up_blocks.3.2.attn.norm_out.weight', 'decoder.up_blocks.3.2.attn.to_k.weight', 'decoder.up_blocks.3.2.attn.to_out.weight', 'decoder.up_blocks.3.2.attn.to_q.weight', 'decoder.up_blocks.3.2.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.3.2.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.3.2.attn.to_v.weight', 'decoder.up_blocks.3.2.conv_out.conv_depth.bias', 'decoder.up_blocks.3.2.conv_out.conv_depth.weight', 'decoder.up_blocks.3.2.conv_out.conv_inverted.bias', 'decoder.up_blocks.3.2.conv_out.conv_inverted.weight', 'decoder.up_blocks.3.2.conv_out.conv_point.weight', 'decoder.up_blocks.3.2.conv_out.norm.bias', 'decoder.up_blocks.3.2.conv_out.norm.weight', 'decoder.up_blocks.3.3.attn.norm_out.bias', 'decoder.up_blocks.3.3.attn.norm_out.weight', 'decoder.up_blocks.3.3.attn.to_k.weight', 'decoder.up_blocks.3.3.attn.to_out.weight', 'decoder.up_blocks.3.3.attn.to_q.weight', 'decoder.up_blocks.3.3.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.3.3.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.3.3.attn.to_v.weight', 'decoder.up_blocks.3.3.conv_out.conv_depth.bias', 'decoder.up_blocks.3.3.conv_out.conv_depth.weight', 'decoder.up_blocks.3.3.conv_out.conv_inverted.bias', 'decoder.up_blocks.3.3.conv_out.conv_inverted.weight', 'decoder.up_blocks.3.3.conv_out.conv_point.weight', 'decoder.up_blocks.3.3.conv_out.norm.bias', 'decoder.up_blocks.3.3.conv_out.norm.weight', 'decoder.up_blocks.4.0.conv.bias', 'decoder.up_blocks.4.0.conv.weight', 'decoder.up_blocks.4.1.attn.norm_out.bias', 'decoder.up_blocks.4.1.attn.norm_out.weight', 'decoder.up_blocks.4.1.attn.to_k.weight', 'decoder.up_blocks.4.1.attn.to_out.weight', 'decoder.up_blocks.4.1.attn.to_q.weight', 'decoder.up_blocks.4.1.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.4.1.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.4.1.attn.to_v.weight', 'decoder.up_blocks.4.1.conv_out.conv_depth.bias', 'decoder.up_blocks.4.1.conv_out.conv_depth.weight', 'decoder.up_blocks.4.1.conv_out.conv_inverted.bias', 'decoder.up_blocks.4.1.conv_out.conv_inverted.weight', 'decoder.up_blocks.4.1.conv_out.conv_point.weight', 'decoder.up_blocks.4.1.conv_out.norm.bias', 'decoder.up_blocks.4.1.conv_out.norm.weight', 'decoder.up_blocks.4.2.attn.norm_out.bias', 'decoder.up_blocks.4.2.attn.norm_out.weight', 'decoder.up_blocks.4.2.attn.to_k.weight', 'decoder.up_blocks.4.2.attn.to_out.weight', 'decoder.up_blocks.4.2.attn.to_q.weight', 'decoder.up_blocks.4.2.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.4.2.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.4.2.attn.to_v.weight', 'decoder.up_blocks.4.2.conv_out.conv_depth.bias', 'decoder.up_blocks.4.2.conv_out.conv_depth.weight', 'decoder.up_blocks.4.2.conv_out.conv_inverted.bias', 'decoder.up_blocks.4.2.conv_out.conv_inverted.weight', 'decoder.up_blocks.4.2.conv_out.conv_point.weight', 'decoder.up_blocks.4.2.conv_out.norm.bias', 'decoder.up_blocks.4.2.conv_out.norm.weight', 'decoder.up_blocks.4.3.attn.norm_out.bias', 'decoder.up_blocks.4.3.attn.norm_out.weight', 'decoder.up_blocks.4.3.attn.to_k.weight', 'decoder.up_blocks.4.3.attn.to_out.weight', 'decoder.up_blocks.4.3.attn.to_q.weight', 'decoder.up_blocks.4.3.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.4.3.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.4.3.attn.to_v.weight', 'decoder.up_blocks.4.3.conv_out.conv_depth.bias', 'decoder.up_blocks.4.3.conv_out.conv_depth.weight', 'decoder.up_blocks.4.3.conv_out.conv_inverted.bias', 'decoder.up_blocks.4.3.conv_out.conv_inverted.weight', 'decoder.up_blocks.4.3.conv_out.conv_point.weight', 'decoder.up_blocks.4.3.conv_out.norm.bias', 'decoder.up_blocks.4.3.conv_out.norm.weight', 'decoder.up_blocks.5.0.attn.norm_out.bias', 'decoder.up_blocks.5.0.attn.norm_out.weight', 'decoder.up_blocks.5.0.attn.to_k.weight', 'decoder.up_blocks.5.0.attn.to_out.weight', 'decoder.up_blocks.5.0.attn.to_q.weight', 'decoder.up_blocks.5.0.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.5.0.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.5.0.attn.to_v.weight', 'decoder.up_blocks.5.0.conv_out.conv_depth.bias', 'decoder.up_blocks.5.0.conv_out.conv_depth.weight', 'decoder.up_blocks.5.0.conv_out.conv_inverted.bias', 'decoder.up_blocks.5.0.conv_out.conv_inverted.weight', 'decoder.up_blocks.5.0.conv_out.conv_point.weight', 'decoder.up_blocks.5.0.conv_out.norm.bias', 'decoder.up_blocks.5.0.conv_out.norm.weight', 'decoder.up_blocks.5.1.attn.norm_out.bias', 'decoder.up_blocks.5.1.attn.norm_out.weight', 'decoder.up_blocks.5.1.attn.to_k.weight', 'decoder.up_blocks.5.1.attn.to_out.weight', 'decoder.up_blocks.5.1.attn.to_q.weight', 'decoder.up_blocks.5.1.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.5.1.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.5.1.attn.to_v.weight', 'decoder.up_blocks.5.1.conv_out.conv_depth.bias', 'decoder.up_blocks.5.1.conv_out.conv_depth.weight', 'decoder.up_blocks.5.1.conv_out.conv_inverted.bias', 'decoder.up_blocks.5.1.conv_out.conv_inverted.weight', 'decoder.up_blocks.5.1.conv_out.conv_point.weight', 'decoder.up_blocks.5.1.conv_out.norm.bias', 'decoder.up_blocks.5.1.conv_out.norm.weight', 'decoder.up_blocks.5.2.attn.norm_out.bias', 'decoder.up_blocks.5.2.attn.norm_out.weight', 'decoder.up_blocks.5.2.attn.to_k.weight', 'decoder.up_blocks.5.2.attn.to_out.weight', 'decoder.up_blocks.5.2.attn.to_q.weight', 'decoder.up_blocks.5.2.attn.to_qkv_multiscale.0.proj_in.weight', 'decoder.up_blocks.5.2.attn.to_qkv_multiscale.0.proj_out.weight', 'decoder.up_blocks.5.2.attn.to_v.weight', 'decoder.up_blocks.5.2.conv_out.conv_depth.bias', 'decoder.up_blocks.5.2.conv_out.conv_depth.weight', 'decoder.up_blocks.5.2.conv_out.conv_inverted.bias', 'decoder.up_blocks.5.2.conv_out.conv_inverted.weight', 'decoder.up_blocks.5.2.conv_out.conv_point.weight', 'decoder.up_blocks.5.2.conv_out.norm.bias', 'decoder.up_blocks.5.2.conv_out.norm.weight']2025-01-03T00:03:28.114182 -
2025-01-03T00:03:29.402325 - !!! Exception during processing !!! The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:03:29.405327 - Traceback (most recent call last):
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 945, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 647, in __getitem__
raise KeyError(key)
KeyError: 'gemma2'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:03:29.408325 - Prompt executed in 4.47 seconds
2025-01-03T00:04:14.853241 - got prompt
2025-01-03T00:04:15.726969 - !!! Exception during processing !!! The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:04:15.729969 - Traceback (most recent call last):
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 945, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 647, in __getitem__
raise KeyError(key)
KeyError: 'gemma2'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:04:15.731969 - Prompt executed in 0.88 seconds
2025-01-03T00:08:55.153890 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-01-03T00:08:55.624720 - [DONE]2025-01-03T00:08:55.624720 -
2025-01-03T00:09:12.246439 -
Fetching done.2025-01-03T00:09:12.246439 -
2025-01-03T00:14:45.570659 - got prompt
2025-01-03T00:14:56.540188 - !!! Exception during processing !!! The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:14:56.543177 - Traceback (most recent call last):
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 945, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 647, in __getitem__
raise KeyError(key)
KeyError: 'gemma2'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_ExtraModels\Gemma\nodes.py", line 67, in load_model
text_encoder_model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=dtype)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\auto_factory.py", line 523, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\transformers\models\auto\configuration_auto.py", line 947, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
2025-01-03T00:14:56.545184 - Prompt executed in 10.97 seconds
Attached Workflow
Please make sure that workflow does not contain any sensitive information such as API keys or passwords.
{"last_node_id":20,"last_link_id":33,"nodes":[{"id":1,"type":"SanaTextEncode","pos":[34.06884765625,-130.40341186523438],"size":[539.967529296875,255.81317138671875],"flags":{},"order":6,"mode":0,"inputs":[{"name":"GEMMA","type":"GEMMA","link":8}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[12,33],"slot_index":0}],"properties":{"Node name for S&R":"SanaTextEncode"},"widgets_values":["A woman with long brown hair and blue eyes, wearing a pink dress and standing in front of a blurred background of a city street."],"color":"#232","bgcolor":"#353"},{"id":2,"type":"SanaResolutionCond","pos":[605.7742919921875,-75.0198745727539],"size":[301.3979187011719,83.60631561279297],"flags":{},"order":8,"mode":0,"inputs":[{"name":"cond","type":"CONDITIONING","link":12},{"name":"width","type":"INT","link":2,"widget":{"name":"width"}},{"name":"height","type":"INT","link":3,"widget":{"name":"height"}}],"outputs":[{"name":"cond","type":"CONDITIONING","links":[13],"slot_index":0}],"properties":{"Node name for S&R":"SanaResolutionCond"},"widgets_values":[1024,1024]},{"id":4,"type":"SanaResolutionSelect","pos":[-334.3977355957031,317.1134338378906],"size":[315,102],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"width","type":"INT","links":[2,4],"slot_index":0},{"name":"height","type":"INT","links":[3,5],"slot_index":1}],"properties":{"Node name for S&R":"SanaResolutionSelect"},"widgets_values":["SanaMS_1600M_P1_D20","1.38"]},{"id":5,"type":"EmptySanaLatentImage","pos":[105.96505737304688,440.70098876953125],"size":[315,106],"flags":{},"order":5,"mode":0,"inputs":[{"name":"width","type":"INT","link":4,"widget":{"name":"width"}},{"name":"height","type":"INT","link":5,"widget":{"name":"height"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[20],"slot_index":0}],"properties":{"Node name for S&R":"EmptySanaLatentImage"},"widgets_values":[512,512,2]},{"id":6,"type":"KSampler","pos":[602.260009765625,73.22354125976562],"size":[315,262],"flags":{},"order":9,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":1},{"name":"positive","type":"CONDITIONING","link":13},{"name":"negative","type":"CONDITIONING","link":11},{"name":"latent_image","type":"LATENT","link":20},{"name":"seed","type":"INT","link":31,"widget":{"name":"seed"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[14,21],"slot_index":0}],"properties":{"Node name for S&R":"KSampler"},"widgets_values":[936979341784666,"randomize",30,3,"euler","normal",1]},{"id":9,"type":"SanaTextEncode","pos":[64.97035217285156,183.81201171875],"size":[449.4744873046875,200],"flags":{},"order":7,"mode":0,"inputs":[{"name":"GEMMA","type":"GEMMA","link":10}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[11,32],"slot_index":0}],"properties":{"Node name for S&R":"SanaTextEncode"},"widgets_values":["blurry, low quality, 3D , cartnoon , extra hands, extra fingers "],"color":"#322","bgcolor":"#533"},{"id":10,"type":"VAEDecode","pos":[699.19384765625,395.64361572265625],"size":[210,46],"flags":{},"order":10,"mode":0,"inputs":[{"name":"samples","type":"LATENT","link":14},{"name":"vae","type":"VAE","link":15}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[16],"slot_index":0}],"properties":{"Node name for S&R":"VAEDecode"},"widgets_values":[]},{"id":11,"type":"ExtraVAELoader","pos":[-416.8797302246094,501.91510009765625],"size":[398.8718566894531,106],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"VAE","type":"VAE","links":[15,23],"slot_index":0}],"properties":{"Node name for S&R":"ExtraVAELoader"},"widgets_values":["Sana_1600M_1024px_vae.safetensors","dcae-f32c32-sana-1.0","FP16"]},{"id":12,"type":"PreviewImage","pos":[1007.628662109375,-122.39791870117188],"size":[1289.3055419921875,936.1663208007812],"flags":{},"order":12,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":16}],"outputs":[],"properties":{"Node name for S&R":"PreviewImage"},"widgets_values":[]},{"id":15,"type":"KSampler","pos":[246.45791625976562,1191.375732421875],"size":[315,262],"flags":{"collapsed":true},"order":11,"mode":2,"inputs":[{"name":"model","type":"MODEL","link":28},{"name":"positive","type":"CONDITIONING","link":33},{"name":"negative","type":"CONDITIONING","link":32},{"name":"latent_image","type":"LATENT","link":21},{"name":"seed","type":"INT","link":30,"widget":{"name":"seed"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[22],"slot_index":0}],"properties":{"Node name for S&R":"KSampler"},"widgets_values":[60991530284536,"randomize",10,2,"euler","normal",0.3]},{"id":16,"type":"VAEDecode","pos":[346.4548034667969,1510.8746337890625],"size":[210,46],"flags":{"collapsed":true},"order":13,"mode":2,"inputs":[{"name":"samples","type":"LATENT","link":22},{"name":"vae","type":"VAE","link":23}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[29],"slot_index":0}],"properties":{"Node name for S&R":"VAEDecode"},"widgets_values":[]},{"id":19,"type":"PreviewImage","pos":[654.5088500976562,1181.4791259765625],"size":[1271.1556396484375,939.3509521484375],"flags":{"collapsed":true},"order":14,"mode":2,"inputs":[{"name":"images","type":"IMAGE","link":29}],"outputs":[],"properties":{"Node name for S&R":"PreviewImage"},"widgets_values":[]},{"id":20,"type":"CR Seed","pos":[106.5981674194336,632.4039306640625],"size":[315,102],"flags":{},"order":2,"mode":0,"inputs":[],"outputs":[{"name":"seed","type":"INT","links":[30,31],"slot_index":0},{"name":"show_help","type":"STRING","links":null}],"properties":{"Node name for S&R":"CR Seed"},"widgets_values":[255093259020808,"randomize"]},{"id":3,"type":"SanaCheckpointLoader","pos":[-364.4580993652344,-42.886985778808594],"size":[330.63165283203125,82],"flags":{},"order":3,"mode":0,"inputs":[],"outputs":[{"name":"model","type":"MODEL","links":[1,28],"slot_index":0}],"properties":{"Node name for S&R":"SanaCheckpointLoader"},"widgets_values":["Sana_1600M_1024px.pth","SanaMS_1600M_P1_D20"]},{"id":8,"type":"GemmaLoader","pos":[-363.2308044433594,123.64581298828125],"size":[349.99859619140625,106],"flags":{},"order":4,"mode":0,"inputs":[],"outputs":[{"name":"GEMMA","type":"GEMMA","links":[8,10],"slot_index":0}],"properties":{"Node name for S&R":"GemmaLoader"},"widgets_values":["unsloth/gemma-2-2b-it-bnb-4bit","cpu","default"]}],"links":[[1,3,0,6,0,"MODEL"],[2,4,0,2,1,"INT"],[3,4,1,2,2,"INT"],[4,4,0,5,0,"INT"],[5,4,1,5,1,"INT"],[8,8,0,1,0,"GEMMA"],[10,8,0,9,0,"GEMMA"],[11,9,0,6,2,"CONDITIONING"],[12,1,0,2,0,"CONDITIONING"],[13,2,0,6,1,"CONDITIONING"],[14,6,0,10,0,"LATENT"],[15,11,0,10,1,"VAE"],[16,10,0,12,0,"IMAGE"],[20,5,0,6,3,"LATENT"],[21,6,0,15,3,"LATENT"],[22,15,0,16,0,"LATENT"],[23,11,0,16,1,"VAE"],[28,3,0,15,0,"MODEL"],[29,16,0,19,0,"IMAGE"],[30,20,0,15,4,"INT"],[31,20,0,6,4,"INT"],[32,9,0,15,2,"CONDITIONING"],[33,1,0,15,1,"CONDITIONING"]],"groups":[{"id":1,"title":"Group","bounding":[192.7167205810547,1017.5338134765625,668.3709716796875,566.7073364257812],"color":"#3f789e","font_size":24,"flags":{}}],"config":{},"extra":{"ds":{"scale":0.8438022155513367,"offset":[668.366620908559,115.06668170315389]},"ue_links":[]},"version":0.4}
Additional Context
(Please add any additional context or steps to reproduce the error here)
Metadata
Metadata
Assignees
Labels
No labels