Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error while using the Skyreel img2vid workflow #371

Open
cassiebug23 opened this issue Feb 18, 2025 · 23 comments
Open

Error while using the Skyreel img2vid workflow #371

cassiebug23 opened this issue Feb 18, 2025 · 23 comments

Comments

@cassiebug23
Copy link

Hi, when I use the skyreel img2vid workflow, I get this error when trying to run it:
!!! Exception during processing !!! Trying to set a tensor of shape torch.Size([3072, 32, 1, 2, 2]) in "weight" (which has shape torch.Size([3072, 16, 1, 2, 2])), this looks incorrect.
Traceback (most recent call last):
File "C:\pinokio\api\comfy.git\app\execution.py", line 327, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "C:\pinokio\api\comfy.git\app\execution.py", line 202, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "C:\pinokio\api\comfy.git\app\execution.py", line 174, in _map_node_over_list
process_inputs(input_dict, i)
File "C:\pinokio\api\comfy.git\app\execution.py", line 163, in process_inputs
results.append(getattr(obj, func)(**inputs))
File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 363, in loadmodel
set_module_tensor_to_device(transformer, name, device=transformer_load_device, dtype=dtype_to_use, value=sd[name])
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\accelerate\utils\modeling.py", line 286, in set_module_tensor_to_device
raise ValueError(
ValueError: Trying to set a tensor of shape torch.Size([3072, 32, 1, 2, 2]) in "weight" (which has shape torch.Size([3072, 16, 1, 2, 2])), this looks incorrect.

What could be causing this error?

@brausesa
Copy link

the same problem

@kwas-123
Copy link

make sure you are using his model and his vae

@cassiebug23
Copy link
Author

make sure you are using his model and his vae

I am. Error still occurs.

@kijai
Copy link
Owner

kijai commented Feb 18, 2025

It looks like the nodes are not up to date for you as it doesn't detect the model type.

@cassiebug23
Copy link
Author

It looks like the nodes are not up to date for you as it doesn't detect the model type.

Updated the nodes but I'm now getting a new error:
!!! Exception during processing !!! Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 64, 40] to have 32 channels, but got 16 channels instead
Traceback (most recent call last):
File "C:\pinokio\api\comfy.git\app\execution.py", line 327, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "C:\pinokio\api\comfy.git\app\execution.py", line 202, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "C:\pinokio\api\comfy.git\app\execution.py", line 174, in _map_node_over_list
process_inputs(input_dict, i)
File "C:\pinokio\api\comfy.git\app\execution.py", line 163, in process_inputs
results.append(getattr(obj, func)(**inputs))
File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 1292, in process
out_latents = model["pipe"](
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\utils_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 776, in call
noise_pred = self.transformer( # For an input image (129, 192, 336) (1, 256, 256)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 952, in forward
img = self.img_in(img)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\embed_layers.py", line 55, in forward
x = self.proj(x)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\conv.py", line 725, in forward
return self._conv_forward(input, self.weight, self.bias)
File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\conv.py", line 720, in _conv_forward
return F.conv3d(
RuntimeError: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 64, 40] to have 32 channels, but got 16 channels instead

@kijai
Copy link
Owner

kijai commented Feb 18, 2025

It looks like the nodes are not up to date for you as it doesn't detect the model type.

Updated the nodes but I'm now getting a new error: !!! Exception during processing !!! Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 64, 40] to have 32 channels, but got 16 channels instead Traceback (most recent call last): File "C:\pinokio\api\comfy.git\app\execution.py", line 327, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\pinokio\api\comfy.git\app\execution.py", line 202, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\pinokio\api\comfy.git\app\execution.py", line 174, in _map_node_over_list process_inputs(input_dict, i) File "C:\pinokio\api\comfy.git\app\execution.py", line 163, in process_inputs results.append(getattr(obj, func)(**inputs)) File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 1292, in process out_latents = model["pipe"]( File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\utils_contextlib.py", line 116, in decorate_context return func(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 776, in call noise_pred = self.transformer( # For an input image (129, 192, 336) (1, 256, 256) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl return forward_call(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 952, in forward img = self.img_in(img) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl return forward_call(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\embed_layers.py", line 55, in forward x = self.proj(x) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl return forward_call(*args, **kwargs) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\conv.py", line 725, in forward return self._conv_forward(input, self.weight, self.bias) File "C:\pinokio\api\comfy.git\app\env\lib\site-packages\torch\nn\modules\conv.py", line 720, in _conv_forward return F.conv3d( RuntimeError: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 64, 40] to have 32 channels, but got 16 channels instead

That would happen if the latent wasn't fed in the image_cond_latents -input, it's possible that since the nodes weren't up to date the workflow got messed up with what goes where, I'd suggest reloading it. I also just updated the whole workflow, it should look like this:

Image

@ShmuelRonen
Copy link

ShmuelRonen commented Feb 18, 2025

I work on WSL2.0 ComfyUI with RTX 4090. after update I got the same error: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 80, 80] to have 32 channels,


ComfyUI Error:

Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: RuntimeError
  • Exception Message: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 80, 80] to have 32 channels, but got 16 channels instead

Stack Trace

  File "/home/sam/ComfyUI/execution.py", line 327, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)

  File "/home/sam/ComfyUI/execution.py", line 202, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)

  File "/home/sam/ComfyUI/execution.py", line 174, in _map_node_over_list
    process_inputs(input_dict, i)

  File "/home/sam/ComfyUI/execution.py", line 163, in process_inputs
    results.append(getattr(obj, func)(**inputs))

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/nodes.py", line 1292, in process
    out_latents = model["pipe"](

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/diffusion/pipelines/pipeline_hunyuan_video.py", line 776, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/modules/models.py", line 952, in forward
    img = self.img_in(img)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/modules/embed_layers.py", line 55, in forward
    x = self.proj(x)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/conv.py", line 608, in forward
    return self._conv_forward(input, self.weight, self.bias)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/conv.py", line 603, in _conv_forward
    return F.conv3d(

Image

@kijai
Copy link
Owner

kijai commented Feb 18, 2025

I work on WSL2.0 ComfyUI with RTX 4090. after update I got the same error: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 80, 80] to have 32 channels,

ComfyUI Error:

Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: RuntimeError
  • Exception Message: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 80, 80] to have 32 channels, but got 16 channels instead

Stack Trace

  File "/home/sam/ComfyUI/execution.py", line 327, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)

  File "/home/sam/ComfyUI/execution.py", line 202, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)

  File "/home/sam/ComfyUI/execution.py", line 174, in _map_node_over_list
    process_inputs(input_dict, i)

  File "/home/sam/ComfyUI/execution.py", line 163, in process_inputs
    results.append(getattr(obj, func)(**inputs))

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/nodes.py", line 1292, in process
    out_latents = model["pipe"](

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/diffusion/pipelines/pipeline_hunyuan_video.py", line 776, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/modules/models.py", line 952, in forward
    img = self.img_in(img)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/modules/embed_layers.py", line 55, in forward
    x = self.proj(x)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/conv.py", line 608, in forward
    return self._conv_forward(input, self.weight, self.bias)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/conv.py", line 603, in _conv_forward
    return F.conv3d(

Image

The input image for this model should be plugged in the image_cond_latents, not samples.

@ShmuelRonen
Copy link

I work on WSL2.0 ComfyUI with RTX 4090. after update I got the same error: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 80, 80] to have 32 channels,

ComfyUI Error:

Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: RuntimeError
  • Exception Message: Given groups=1, weight of size [3072, 32, 1, 2, 2], expected input[2, 16, 25, 80, 80] to have 32 channels, but got 16 channels instead

Stack Trace

  File "/home/sam/ComfyUI/execution.py", line 327, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)

  File "/home/sam/ComfyUI/execution.py", line 202, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)

  File "/home/sam/ComfyUI/execution.py", line 174, in _map_node_over_list
    process_inputs(input_dict, i)

  File "/home/sam/ComfyUI/execution.py", line 163, in process_inputs
    results.append(getattr(obj, func)(**inputs))

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/nodes.py", line 1292, in process
    out_latents = model["pipe"](

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/diffusion/pipelines/pipeline_hunyuan_video.py", line 776, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/modules/models.py", line 952, in forward
    img = self.img_in(img)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/custom_nodes/ComfyUI-HunyuanVideoWrapper/hyvideo/modules/embed_layers.py", line 55, in forward
    x = self.proj(x)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
    return forward_call(*args, **kwargs)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/conv.py", line 608, in forward
    return self._conv_forward(input, self.weight, self.bias)

  File "/home/sam/ComfyUI/venv/lib/python3.10/site-packages/torch/nn/modules/conv.py", line 603, in _conv_forward
    return F.conv3d(

Image

The input image for this model should be plugged in the image_cond_latents, not samples.


Thanks.
now raise a different error: HyVideoSampler
backend='inductor' raised:
CompilationError: at 8:11:
def triton_poi_fused__to_copy_1(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr):
xnumel = 56623104
xoffset = tl.program_id(0) * XBLOCK
xindex = xoffset + tl.arange(0, XBLOCK)[:]
xmask = xindex < xnumel
x0 = xindex
tmp0 = tl.load(in_ptr0 + (x0), None)
tmp1 = tmp0.to(tl.float32)
^

Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information

You can suppress this exception and fall back to eager by setting:
import torch._dynamo
torch._dynamo.config.suppress_errors = True

@kijai
Copy link
Owner

kijai commented Feb 18, 2025

Torch compile errors are harder to troubleshoot, can't really help there

@ShmuelRonen
Copy link

Torch compile errors are harder to troubleshoot, can't really help there

So.... I must wait that ComfyUI will bring a lighter version as they done to Hunyuan t2v

@kijai
Copy link
Owner

kijai commented Feb 18, 2025

Torch compile errors are harder to troubleshoot, can't really help there

So.... I must wait that ComfyUI will bring a lighter version as they done to Hunyuan t2v

Compile isn't a necessity, it's a bonus, and the error isn't specific to these nodes and you'd face same issue using compile with the native nodes.

@ShmuelRonen
Copy link

Torch compile errors are harder to troubleshoot, can't really help there

So.... I must wait that ComfyUI will bring a lighter version as they done to Hunyuan t2v

Compile isn't a necessity, it's a bonus, and the error isn't specific to these nodes and you'd face same issue using compile with the native nodes.

This configuration works for me:

Image

@jtsanborn1
Copy link

In my case i'm getting this: The size of tensor a (32) must match the size of tensor b (16) at non-singleton dimension 1

Any sight? @kijai
Image

Image

](url)

@kijai
Copy link
Owner

kijai commented Feb 19, 2025

In my case i'm getting this: The size of tensor a (32) must match the size of tensor b (16) at non-singleton dimension 1

Any sight? @kijai Image

Image

](url)

Don't put anything in the samples input, it's for video2video processes only (and leapfusion as it uses similar approach).

@jtsanborn1
Copy link

@kijai Thanks for your reply. Now the output is completely black. Any sight why is this happening? No matter what I do, it doesnt work. I lower the number of frames just to try different settings to see if works but it just doesnt work.
WIN 10 RTX 3090

Image

@kijai
Copy link
Owner

kijai commented Feb 19, 2025

@kijai Thanks for your reply. Now the output is completely black. Any sight why is this happening? No matter what I do, it doesnt work. I lower the number of frames just to try different settings to see if works but it just doesnt work.
WIN 10 RTX 3090

Image

Which version of pytorch are you using? This is known issue with versions prior to 2.5.0.

@jtsanborn1
Copy link

pytorch version: 2.3.1+cu121

@jtsanborn1
Copy link

I just updated the pytorch version: 2.5.1+cu121, also tried with 2.5.1+cu124
But the output is just a random noise...

Image

@kijai
Copy link
Owner

kijai commented Feb 19, 2025

I just updated the pytorch version: 2.5.1+cu121, also tried with 2.5.1+cu124 But the output is just a random noise...

Image

Hmm... that I don't have experience with, try with minimum of 17 frames and without the enhance a video.

@jtsanborn1
Copy link

@kijai I can't thank you enough!! I was getting crazy as I spend two days trying to make it work and nothing worked, but it is working now! Thanks!

@Sarley-Can
Copy link

Failed to import transformers.models.timm_wrapper.configuration_timm_wrapper because of the following error (look up to see its traceback):
cannot import name 'ImageNetInfo' from 'timm.data' (C:\Users\TR\Documents\ComfyUI\custom_nodes\ComfyUI-tbox\src\timm\data_init_.py)

Hello there i am getting this issue while trying your workflow any ideas @kijai ?

@kijai
Copy link
Owner

kijai commented Feb 21, 2025

Failed to import transformers.models.timm_wrapper.configuration_timm_wrapper because of the following error (look up to see its traceback): cannot import name 'ImageNetInfo' from 'timm.data' (C:\Users\TR\Documents\ComfyUI\custom_nodes\ComfyUI-tbox\src\timm\data__init__.py)

Hello there i am getting this issue while trying your workflow any ideas @kijai ?

That error would come from having too old transformers and/or timm versions, updating those should fix it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants