▲ | MintsJohn 6 days ago | ||||||||||||||||
This is what finetuning has been all about since stable diffusion 1.5 and especially SDXL. And even something StabilityAI base models excelled at in the open weights category. (Midjourney has always been the champion, but proprietary) Sadly with SAI going effectively bankrupt things changed, their rushed 3.0 model was broken beyond repair and the later 3.5 just unfinished or something (the api version is remarkably better), gens full of errors and artifacts even though the good ones looked great. It turned out hard to finetune as well. In the mean time flux got released, but that model can be fried (as in one concept trained in) but not finetuned (this krea flux is not based on the open weights flux). Add to that that as models got bigger training/finetuning now costs an arm and a leg, so here we are, a year after flux got released a good finetune is celebrated as the next new thing :) | |||||||||||||||||
▲ | vunderba 6 days ago | parent | next [-] | ||||||||||||||||
Agreed. From the article: > Model builders have been mostly focused on correctness, not aesthetics. Researchers have been overly focused on the extra fingers problem. While that might be true for the foundational models - the author seems to be neglecting the tens of thousands of custom LoRAs to customize the look of an image. > Users fight the “AI Look” with heavy prompting and even fine-tuning IMHO it is significantly easier to fix an aesthetic issue than an adherence issue. You can take a poor quality image, use ESRGAN upscalers, img2img using it as a ControlNet, run it through a different model, add LoRAs, etc. I have done some nominal tests with Krea but mostly around adherence. I'd be curious to know if they've reduced the omnipresent bokeh / shallow depth of field given that it is Flux based. | |||||||||||||||||
| |||||||||||||||||
▲ | dvrp 5 days ago | parent | prev [-] | ||||||||||||||||
It is not just a fine-tune. |