Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature request: Switch off Unet for DiT #11

Open
moiseshorta opened this issue Feb 6, 2025 · 6 comments
Open

Feature request: Switch off Unet for DiT #11

moiseshorta opened this issue Feb 6, 2025 · 6 comments

Comments

@moiseshorta
Copy link

Hello,

I've been reading a lot of the SOTA papers on audio and video generation using Rectified Flows, and it seems most are using Transformers instead of Unets.

Are there any plans to implement such an architecture change? They seem to improve greatly in performance, as in this implementation: https://github.com/cloneofsimo/minRF

Would be great to see it here, as it's a very clear to understand codebase, thanks again for opensourcing it!

@lucidrains
Copy link
Owner

@moiseshorta yes you are correct, pure attention has basically completely taken over

why not use Simo's implementation instead of the one here? is there anything lacking in his? he's a pretty amazing guy in general

@moiseshorta
Copy link
Author

As I mentioned, your implementation seems a bit more clear to me than Simo's. Although I will give his a try as well :)

@lucidrains
Copy link
Owner

ohh, ok, I've seen him live code before and was very impressed. but yeah sure i can add DiT here, or perhaps just make it x-transformer compatible

@moiseshorta
Copy link
Author

yeah, that would be great. I'm currently trying to scale up your implementation but I seem to be either very quickly overfitting or just NaN/Inf gradients while training on a bigger dataset...

@lucidrains
Copy link
Owner

@moiseshorta overfitting vs divergence are two very different things, and there are tricks of the trade for both of those issues.

sounds good, give me some time, too many projects but i do want to flesh out this repo a bit more, or dog food other libraries

@moiseshorta
Copy link
Author

Yes, definitely. On some occasions there is overfitting, tried mitigating with dropout. On some occasions the divergence has been mitigated with lower LR. Any other advice is greatly appreciated! Looking forward to the implementation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants