How to run WanAnimatePipeline with GGUF quantization?

#2
by awssamdwar - opened

Hello,

I tried to run WanAnimatePipeline and WanAnimateTransformer3DModel with GGUF Quantized models, but getting operation unsupported for WanAnimateTransformer3DModel. Any pointers on how to run Wan Animate with less GPU memory? currently running out of GPU memory when trying to run the 1280 x 780 example

I have been looking to create a state_dict conversion function, so far so good until I get to the motion_encoder weights, I am stuck at the mapping, can't make sense which is which. Would be great if someone can shed a light on how transform weights maps to diffuser weights for the motion_encoder piece. In the Wan Animate paper it is called "Body Adapter" if I am not mistaken, but not much mentioned about it.

@dg439 would be great if you can shed a light on the above

Finally, I was able to create a conversion function, will publish on diffusers git soon

https://github.com/huggingface/diffusers/pull/12691 , although I can load the GGUF files now, I am getting errors related to dtype mismatch when a matrix operation is performed πŸ˜₯

Was able to fix the issue and added fix to same pull request

Sign up or log in to comment