-
Notifications
You must be signed in to change notification settings - Fork 892
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for Prodigy(Dadapt variety for Dylora) #585
Conversation
Thanks, looks very good! I will check it out when I have time! |
Works well for LORAs on SDXL. Convergence to the optimal LR can be a bit slow (1000 steps) compared to DAdapt, or maybe it's just SDXL being big. Does not blow up compared to DAdapt though. Needs more testing. Question: which version was implemented, Prodigy (2 in the paper) or Resetting (3)? |
I would suggest modifying the value of d0 to accommodate SDXL(5e-7) as well as dylora(5e-4), which are models that require a larger initial learning rate. |
how do you use the generated safetensors file? can you use with diffusers pipeline? |
Thanks for the tutorial, good stuff. Could use a straightforward way to set d0 (the initial LR) if I know that the algorithm will just "waste time" at 1e-6 in the beginning. Edit nvm I see, just add d0=(number) and d_coef=(number) to the args. |
The LORA pipeline works with ComfyUI at the moment. Don't know about other impl. |
official only 2,no resetting code
|
can you share example json file please |
Something like this? This uses the "new" refiner workflow. I assume you can also train a LORA for the refiner, but unsure of the purpose or how to use it, or what sort of training set would you use. Thus, this is also likely not the correct final workflow. |
hlelo so many parameters are missing here can you share full command like below? I tried like this it didnt work
|
Details see here↓
facebookresearch/dadaptation#24
Paper: https://arxiv.org/pdf/2306.06101.pdf
official repository:https://github.com/konstmish/prodigy