Could it be possible to run Pixtral-12b in the Comfy UI ? #4899
-
https://huggingface.co/mistral-community/pixtral-12b-240910 More details here: https://www.youtube.com/watch?v=PfzPfB3esG4 Is it technically possible to have Comfy UI support ( native or via custom node ) ? Could it give better images or have other capabilities beyond Flux ? |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 4 replies
-
Yes, similar models like Qwen2-VL *) and MiniCPM already have custom nodes. I don't think Pixtral is revolutionary in any way, just another late-fusion multimodal model that can interpret images, but output text only. Early-fusion models that can understand image and text and can also output image and text such as Meta's Chameleon and Transfusion is what's really exciting imo. *) https://github.com/IuvenisSapiens/ComfyUI_Qwen2-VL-Instruct |
Beta Was this translation helpful? Give feedback.
-
Yes: https://github.com/ShmuelRonen/ComfyUI_pixtral_vision |
Beta Was this translation helpful? Give feedback.
-
Is there a node that runs it locally instead of api? |
Beta Was this translation helpful? Give feedback.
Yes:
https://github.com/ShmuelRonen/ComfyUI_pixtral_vision