-
I follow the instruction to finetune 512px and it works well. However, When switch to 1024px-MS version, it doesn't produce reasonable result. Is it possible to share a official implementation to finetune 1024px-MS version of your work? This will be really helpful! |
Beta Was this translation helpful? Give feedback.
Replies: 6 comments 3 replies
-
Beta Was this translation helpful? Give feedback.
-
May be there is something wrong with learnable scale and ratio embedding? Should I keep them fixed during training? |
Beta Was this translation helpful? Give feedback.
-
@SunzeY Same situation; I'm waiting for the author's response, too. |
Beta Was this translation helpful? Give feedback.
-
I think this is caused by the wrong positional embedding. Set the lewei_scale to 2.0 in the config file and have a try! |
Beta Was this translation helpful? Give feedback.
-
In inference.py, no-ema checkpoint is loaded, while fine-tuning ema checkpoint is saved in 'state_dict_ema' key. Replace with the following code solved my problem. PixArt-alpha/scripts/inference.py Line 158 in 82c8559 |
Beta Was this translation helpful? Give feedback.
-
Does your config continue from the 1024 checkpoint? I had that issue. |
Beta Was this translation helpful? Give feedback.
I think this is caused by the wrong positional embedding. Set the lewei_scale to 2.0 in the config file and have a try!