Potential for simple video transfer [Demo showcase] #447
Replies: 3 comments 3 replies
-
|
Ah I can't see your video on my side... For extremely quick textual inversion, keep an eye on this: https://arxiv.org/pdf/2302.13848.pdf (paper reports ~0.05 sec using a single image) |
Beta Was this translation helpful? Give feedback.
-
|
Stay tuned, new full feature showcase of very very simple control net video (simple batch, no feedback) is rendering. Full video: https://www.youtube.com/watch?v=-B0h254Gn1A Video preview added, omg, how is this possible: |
Beta Was this translation helpful? Give feedback.
-
|
Improved version of the pipe in 2x the FPS and 2x the res using additonal LORA to increase stability. Includes some pretty good post processing. If you preprocess externally you get much better results as the annotators are kind of bad compared to the current models available in repos. |
Beta Was this translation helpful? Give feedback.





Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
spent last 2-3 weeks figuring out how this all works and putting some real simple stuff together into a pipeline so videos with reasonable stability can be produced - I think we reached that point. ControlNet was the final piece (perfect timing), all in all its a combination of ~8-10 free tools/AIs.
Following video is produced from ~80 unrelated source images used to train model - hope to reduce this to 1 source image, with user guided expansion to <10 ai generated supplemental sources. Made on a trained model ~aprx 8 hours - hope to reduce under 30m when I get the hang of it. Input for the video is unrelated 3D model animation that looks bit of familiar, but not really close as I made the model in like 5 minutes. Theres tons of supplemental data pregenerated and the generation is basically automated and only requires the guidance video and 1 seed frame - the rest of the conversion is completely unguided.
Hope to improve this in the following weeks including automation of the fine tuning, theres a ton of adjustments done to the init params so it 'works'. Note its generated in full 30 fps to showcase 'stability' and then interpolated to 60. If it was done in 5fps with human oversight the result would ofc be better, but that is not the point.
ex11-2x-RIFE-RIFE4.0-60fps.mp4
Thanks for the amazing tool. Feel free to use this for promo.
Beta Was this translation helpful? Give feedback.
All reactions