SD 2.1 Text Encoder in 256 conv LyCORIS #565
idlebg
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Enhance your model's quality and sharpness using your own pre-trained Unet.
Trained on "121361" images.
How is it working for your 2.1? :)
Just uploaded it to Civit https://civitai.com/models/83622/difusionai-text-encoder-sd-21-lycoris
The text encoder (without UNET) is wrapped in LyCORIS. Optimizer: torch.optim.adamw.AdamW(weight_decay=0.01, betas=(0.9, 0.99))
Network dimension/rank: 768.0 Alpha: 768.0 Module: lycoris.kohya {'conv_dim': '256', 'conv_alpha': '256', 'algo': 'loha'}
Large size due to Lyco CONV 256 .. and saved in Float :D
For a1111
Install https://github.com/KohakuBlueleaf/a1111-sd-webui-lycoris
Download di.ffusion.ai-tXe-FXAA to /models/Lycoris
Option1:
Insert lyco:[di.FFUSION.ai](http://di.ffusion.ai/)-tXe-FXAA:1.0 to prompt
No need to split Unet and Text Enc as its only TX encoder there.
You can go up to 2x weights
Option2: If you need it always ON (ex run a batch from txt file) then you can go to settings / Quicksettings list
add sd_lyco
restart and you should have a drop-down now 🤟 🥃

More info:
"ss_text_encoder_lr": "1e-07",
"ss_keep_tokens": "3",
"ss_network_args": {
"conv_dim": "256",
"conv_alpha": "256",
"algo": "loha"
},
"img_count": 121361
}
"ss_total_batch_size": "100",
"ss_network_dim": "768",
"ss_max_bucket_reso": "1024",
"ss_network_alpha": "768.0",
"ss_steps": "2444",
"sshs_legacy_hash": "539b2745",
"ss_batch_size_per_device": "20",
"ss_max_train_steps": "2444",
"ss_network_module": "lycoris.kohya",
This is a heavy experimental version we used to test even with sloppy captions (quick WD tags and terrible clip), yet the results were satisfying.
Note: This is not the text encoder used in the official FFUSION AI model.
Beta Was this translation helpful? Give feedback.
All reactions