Problem
The XLabs LoadFluxIPAdapter node (from x-flux-comfyui) requires a 768-dim CLIP ViT-L/14 vision encoder. cog-comfyui currently ships clip-vit-large-patch14.bin in supported weights, but PyTorch 2.6 rejects .bin (pickle) files with weights_only=True:
_pickle.UnpicklingError: Weights only load failed. In PyTorch 2.6, we changed the
default value of the `weights_only` argument in `torch.load` from `False` to `True`.
WeightsUnpickler error: Unsupported operand 168
The code path is:
x-flux-comfyui/nodes.py:517 → load_clip_vision(path_clip)
→ comfy/clip_vision.py:160 → torch.load(ckpt, weights_only=True)
The only .safetensors CLIP vision models available (clip_vision_h.safetensors, IPAdapter_image_encoder_sd15.safetensors) output 1024-dim embeddings, causing a dimension mismatch:
RuntimeError: mat1 and mat2 shapes cannot be multiplied (1x1024 and 768x16384)
Request
Please add the safetensors version of openai/clip-vit-large-patch14 to supported weights:
This is the safetensors equivalent of the clip-vit-large-patch14.bin you already host — same model, safe format that works with PyTorch 2.6+.
Impact
This unblocks the XLabs Flux IP-Adapter (both v1 and v2) for all cog-comfyui users. IP-Adapter is essential for image-conditioned generation (e.g., clothing/product transfer in e-commerce photography). Currently the only workaround is to strip IP-Adapter from the workflow entirely.
Environment
- cog-comfyui with
x-flux-comfyui custom node
flux-ip-adapter.safetensors (v1) or flux-ip-adapter-v2.safetensors (v2) in xlabs/ipadapters/
- PyTorch 2.6+ (the version in current cog-comfyui builds)
Reproduction
- Create a workflow using
LoadFluxIPAdapter + ApplyFluxIPAdapter nodes
- Set
clip_vision to clip-vit-large-patch14.bin → fails with pickle error
- Set
clip_vision to clip_vision_h.safetensors → fails with dimension mismatch (1024 vs 768)
No available weight in the current supported list provides a 768-dim .safetensors CLIP vision encoder.
Thank you!
Problem
The XLabs
LoadFluxIPAdapternode (fromx-flux-comfyui) requires a 768-dim CLIP ViT-L/14 vision encoder. cog-comfyui currently shipsclip-vit-large-patch14.binin supported weights, but PyTorch 2.6 rejects.bin(pickle) files withweights_only=True:The code path is:
The only
.safetensorsCLIP vision models available (clip_vision_h.safetensors,IPAdapter_image_encoder_sd15.safetensors) output 1024-dim embeddings, causing a dimension mismatch:Request
Please add the safetensors version of
openai/clip-vit-large-patch14to supported weights:model.safetensors(1.71 GB)clip-vit-large-patch14.safetensors(inComfyUI/models/clip_vision/)This is the safetensors equivalent of the
clip-vit-large-patch14.binyou already host — same model, safe format that works with PyTorch 2.6+.Impact
This unblocks the XLabs Flux IP-Adapter (both v1 and v2) for all cog-comfyui users. IP-Adapter is essential for image-conditioned generation (e.g., clothing/product transfer in e-commerce photography). Currently the only workaround is to strip IP-Adapter from the workflow entirely.
Environment
x-flux-comfyuicustom nodeflux-ip-adapter.safetensors(v1) orflux-ip-adapter-v2.safetensors(v2) inxlabs/ipadapters/Reproduction
LoadFluxIPAdapter+ApplyFluxIPAdapternodesclip_visiontoclip-vit-large-patch14.bin→ fails with pickle errorclip_visiontoclip_vision_h.safetensors→ fails with dimension mismatch (1024 vs 768)No available weight in the current supported list provides a 768-dim
.safetensorsCLIP vision encoder.Thank you!