Replies: 2 comments
-
I would like to know this as well. Seems like what you have pointed out is all valid point. |
Beta Was this translation helpful? Give feedback.
-
I guess one possible reason is that as a defined part of the standard diffusers implementation, you can just save the model and load it and it will work out of the box without having to patch the pipe. But that seems to negate two things that I found so cool about LoRA: 1) the convenience of tiny files storing just the LoRA weights (with the minor inconvenience of having to patch them into the standard model), and 2) the fun of just applying LoRA in all sorts of different places and observing what happens (without having to rewrite classes to make it work). If I could convince myself (or someone could clue me in) that the math works out the same, then I would definitely stick with the way you've set it up here. |
Beta Was this translation helpful? Give feedback.
-
There is a WIP PR huggingface/diffusers#1884 that is implementing LoRA in the diffusers library. Can someone tracking this explain why one has to alter the CrossAttention classes?
@yasyf , I see you're working on a FLAX version, any insight here?
Thanks much in advance.
Beta Was this translation helpful? Give feedback.
All reactions