Just wondering, this has been a head-scratcher for me for a while.
Everywhere I look claims DoRA is superior to LoRA in what seems like all aspects. It doesn’t require more power or resources to train.
I googled DoRA training for newer models – Wan, Qwen, etc. Didn’t find anything, except a reddit post from a year ago asking pretty much exactly what I’m asking here today lol. And every comment seems to agree DoRA is superior. And Comfy has supported DoRA now for a long time.
Yet, here we are – still training LoRAs when there’s been a better option for years? This community is always fairly quick to adopt the latest and greatest. It’s odd this slipped through? I use diffusion-pipe to train pretty much everything now. I’m curious to know if theres a way I could train DoRAs with that. Or if there is a different method out there right now that is capable of training a wan DoRA.
Thanks for any insight, and curious to hear others opinions on this.
Edit: very insightful and interesting responses, my opinion has definitely shifted. @roger_ducky has a great explanation of DoRA drawbacks I was unaware of. Also cool to hear from people who had worse results than LoRA training using the same dataset/params. It sounds like sometimes LoRA is better, and sometimes DoRA is better, but DoRA is certainly not better in every instance – as I was initially led to believe. But still feels like DoRAs deserve more exploration and testing than they’ve had, especially with newer models.
submitted by /u/Realistic_Rabbit5429
[link] [comments]
submitted by /u/Murky_Foundation5528 [link] [comments]
Multi-tool-integrated reasoning enables LLM-empowered tool-use agents to solve complex tasks by interleaving natural-language reasoning with…
Saish Sali, Nipun Kumar, Sura ElamuruguIntroductionAs Netflix has grown, machine learning continues to support our…
Business leaders across industries rely on operational dashboards as the shared source of truth that…
OpenAI’s cofounder and president revealed in federal court on Monday that he’s one of the…