not working

#3
by deeduckme - opened

[ComfyUI-Manager] All startup tasks have been completed.
got prompt
model weight dtype torch.bfloat16, manual cast: None
model_type FLUX
unet unexpected: ['audio_embeddings_connector.learnable_registers', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.k_norm.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.q_norm.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_k.bias', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_k.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_out.0.bias', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_out.0.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_q.bias', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_q.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_v.bias', 'audio_embeddings_connector.transformer_1d_blocks.0.attn1.to_v.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.ff.net.0.proj.bias', 'audio_embeddings_connector.transformer_1d_blocks.0.ff.net.0.proj.weight', 'audio_embeddings_connector.transformer_1d_blocks.0.ff.net.2.bias', 'audio_embeddings_connector.transformer_1d_blocks.0.ff.net.2.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.k_norm.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.q_norm.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_k.bias', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_k.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_out.0.bias', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_out.0.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_q.bias', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_q.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_v.bias', 'audio_embeddings_connector.transformer_1d_blocks.1.attn1.to_v.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.ff.net.0.proj.bias', 'audio_embeddings_connector.transformer_1d_blocks.1.ff.net.0.proj.weight', 'audio_embeddings_connector.transformer_1d_blocks.1.ff.net.2.bias', 'audio_embeddings_connector.transformer_1d_blocks.1.ff.net.2.weight', 'video_embeddings_connector.learnable_registers', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.k_norm.weight', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.q_norm.weight', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_k.bias', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_k.weight', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_out.0.bias', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_out.0.weight', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_q.bias', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_q.weight', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_v.bias', 'video_embeddings_connector.transformer_1d_blocks.0.attn1.to_v.weight', 'video_embeddings_connector.transformer_1d_blocks.0.ff.net.0.proj.bias', 'video_embeddings_connector.transformer_1d_blocks.0.ff.net.0.proj.weight', 'video_embeddings_connector.transformer_1d_blocks.0.ff.net.2.bias', 'video_embeddings_connector.transformer_1d_blocks.0.ff.net.2.weight', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.k_norm.weight', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.q_norm.weight', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_k.bias', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_k.weight', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_out.0.bias', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_out.0.weight', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_q.bias', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_q.weight', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_v.bias', 'video_embeddings_connector.transformer_1d_blocks.1.attn1.to_v.weight', 'video_embeddings_connector.transformer_1d_blocks.1.ff.net.0.proj.bias', 'video_embeddings_connector.transformer_1d_blocks.1.ff.net.0.proj.weight', 'video_embeddings_connector.transformer_1d_blocks.1.ff.net.2.bias', 'video_embeddings_connector.transformer_1d_blocks.1.ff.net.2.weight']
VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16
no CLIP/text encoder weights in checkpoint, the text encoder model will not be loaded.
Requested to load LTXAVTEModel_
loaded completely; 25965.49 MB loaded, full load: True
CLIP/text encoder model load device: cpu, offload device: cpu, current: cpu, dtype: torch.float16
Requested to load VideoVAE
0 models unloaded.
loaded partially; 0.00 MB usable, 0.00 MB loaded, 2378.23 MB offloaded, 648.02 MB buffer reserved, lowvram patches: 0
Using sage attention mode: auto
Requested to load LTXAV
0 models unloaded.
loaded partially; 0.00 MB usable, 0.00 MB loaded, 35997.23 MB offloaded, 448.11 MB buffer reserved, lowvram patches: 0
100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 8/8 [18:39<00:00, 140.00s/it]
Requested to load AudioVAE
loaded completely; 1857.86 MB usable, 415.20 MB loaded, full load: True
Requested to load LTXAV
0 models unloaded.
loaded partially; 0.00 MB usable, 0.00 MB loaded, 35997.23 MB offloaded, 704.09 MB buffer reserved, lowvram patches: 1344
100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 8/8 [12:19<00:00, 92.40s/it]
Requested to load VideoVAE
0 models unloaded.
loaded partially; 0.00 MB usable, 0.00 MB loaded, 2378.23 MB offloaded, 648.02 MB buffer reserved, lowvram patches: 0
!!! Exception during processing !!! cannot open resource
Traceback (most recent call last):
File "/home/mysparkmachine/ComfyUI/execution.py", line 518, in execute
output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/mysparkmachine/ComfyUI/execution.py", line 329, in get_output_data
return_values = await _async_map_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/mysparkmachine/ComfyUI/execution.py", line 303, in _async_map_node_over_list
await process_inputs(input_dict, i)
File "/home/mysparkmachine/ComfyUI/execution.py", line 291, in process_inputs
result = f(**inputs)
^^^^^^^^^^^
File "/home/mysparkmachine/ComfyUI/custom_nodes/comfyui_fill-nodes/nodes/image/FL_ImageNotes.py", line 32, in add_notes
result_img = self.add_text_bar(img, text, bar_height, text_size, border, word_wrap, bar_position)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/mysparkmachine/ComfyUI/custom_nodes/comfyui_fill-nodes/nodes/image/FL_ImageNotes.py", line 48, in add_text_bar
font = ImageFont.truetype(font_path, text_size)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/mysparkmachine/comfyui-env/lib/python3.12/site-packages/PIL/ImageFont.py", line 880, in truetype
return freetype(font)
^^^^^^^^^^^^^^
File "/home/mysparkmachine/comfyui-env/lib/python3.12/site-packages/PIL/ImageFont.py", line 877, in freetype
return FreeTypeFont(font, size, index, encoding, layout_engine)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/mysparkmachine/comfyui-env/lib/python3.12/site-packages/PIL/ImageFont.py", line 285, in init
self.font = core.getfont(
^^^^^^^^^^^^^
OSError: cannot open resource

The error seems to be here:

"no CLIP/text encoder weights in checkpoint, the text encoder model will not be loaded."

No CLIP, no generation.

While you have:
"CLIP/text encoder model load device: cpu, offload device: cpu, current: cpu, dtype: torch.float16"
it doesn't mean it's set at the right place, otherwise the "will not be loaded" message wouldn't appear.

Sign up or log in to comment