[-] Even_Adder@lemmy.dbzer0.com 4 points 3 hours ago

You can never learn anything with these clickbait headlines.

[-] Even_Adder@lemmy.dbzer0.com 3 points 1 day ago

You're good to go as long as it's original enough. If it isn't then that's just copyright infringement. You might as well have right-clicked "save as".

[-] Even_Adder@lemmy.dbzer0.com 13 points 1 day ago

This isn't always true.

[-] Even_Adder@lemmy.dbzer0.com 3 points 3 days ago

Why do you have it?

[-] Even_Adder@lemmy.dbzer0.com 7 points 3 days ago

Or just not show people what you're typing.

[-] Even_Adder@lemmy.dbzer0.com 3 points 5 days ago

They're not wearing the eye protection.

[-] Even_Adder@lemmy.dbzer0.com 24 points 5 days ago

I spotted that too. It has to be.

1
1
1

Among other improvements, the new defaults set --flux_guidance_value=1, removing the need to use CFG nodes at inference, reducing generation time and improving image quality of LoRAs slightly.

Changelog: https://github.com/bghira/SimpleTuner/releases/tag/v0.9.8.1

Sample LoRA: https://huggingface.co/ptx0/flux-dreambooth-lora-r16-dev-cfg1/blob/main/pytorch_lora_weights.safetensors

1
1
1
1

From The Hugging Face Model Card:

Not Ready

This is a WIP and not ready for use. This is an early testing version for research and development. You may know what this is and how to use it, if so, feel free, but it will change as I continue to develop it. I plan to do many updates to it frequently. So you may want to set a revision if you intend to use it anyway.

What is this?

FLUX.1-schnell is an amazing distilled model with an apache 2.0 license. However, it is not finetunable. LoRAs, IP adapters, control nets, etc, cannot be trained on it because it is distilled. The goal of this project is to finetune a non-distilled version of it that can be used as a training base to train adapters for FLUX.1-schnell.

Current Issues

Since we are breaking the distillation, this model will need many steps and guidance to produce good results. Currently, this model, like the schnell version, does not have guidance embeddings. Because of this (and possible other factors) images generated with this model will not look great. However, this hopefully will not affect training, since guidance is not used during training. The things trained on this model are intended to be used on the schnell version anyway. I am working on training guidance embeddings for it, but hopefully it will work as a training base without them.

1

Quoted From Reddit:

Release: https://github.com/bghira/SimpleTuner/releases/tag/v0.9.8

It's here! Runs on 24G cards using Quanto's 8bit quantisation or down to 13G with a 2bit base model for the truly terrifying potato LoRA of your dreams!

If you're after accuracy, a 40G card will do Just Fine, with 80G cards being somewhat of a sweet spot for larger training efforts.

What you get:

  • LoRA, full tuning (but probably just don't do that)
  • Documentation to get you started fast
  • Probably better for just square crop training for now - might artifact for weird resolutions
  • Quantised base model unlocks the ability to safely use Adafactor, Prodigy, and other neat optimisers as a consolation prize for losing access to full bf16 training (AdamWBF16 just won't work with Quanto)

not a fine-tune, but, Flux-fast

frequently observed questions

  • 10k images isn't a requirement for training, that's just a healthy amount of regularisation data to have.

  • Regularisation data with text in it is needed to retain text while tuning Flux. It's sensitive to forgetting.

  • you can finetune either dev or schnell, and you probably don't even need special training dynamics for schnell. it seems to work just fine, but at lower quality than dev, because the base model is lower quality.

  • yes, multiple 4090s or 3090s can be used. no, it's probably not a good idea to try splitting the model across them - stick with quantising and LoRAs.

thank you

You all had a really good response to my work; as well as respect for the limitations of the progress at that point, and the optimism on what can happen next.

I'm not sure whether we can really "improve" this state of the art model - probably merely being able to change it without ruining it is good enough for me.

further work, help needed

If any of you would like to take on any of the items in this issue, we can implement them into SimpleTuner next and unlock another level of fine-tuning efficiency: https://github.com/huggingface/peft/issues/1935

The principle improvement for Flux here will be the ability to train quantised LoKr models, where even the weights of the LoRA itself become quantised in addition to the base model.

1
1
1
1
submitted 1 month ago* (last edited 1 month ago) by Even_Adder@lemmy.dbzer0.com to c/stable_diffusion@lemmy.dbzer0.com

Github user bghira discovered both Schnell and Dev are distilled from Black Forest Labs' Pro model and probably won't be traditionally tuneable.

[-] Even_Adder@lemmy.dbzer0.com 152 points 7 months ago

I always thought it was a man at the gallows, not someone killing themselves.

[-] Even_Adder@lemmy.dbzer0.com 223 points 7 months ago

Reminder that this is made by Ben Zhao, the University of Chicago professor who stole open source code for his last data poisoning scheme.

[-] Even_Adder@lemmy.dbzer0.com 227 points 10 months ago

You're not going to develop AI for the benefit of humanity at Microsoft. If they go there, we'll know "Open"AI's mission was all a lie.

[-] Even_Adder@lemmy.dbzer0.com 106 points 11 months ago

Tenuous at best.

view more: next ›

Even_Adder

joined 1 year ago