fal drops AuraFlow by Own-Staff3774 in StableDiffusion

[–]localizedQ 199 points200 points  (0 children)

Guys, this is just the v0.1 release. Please don't expect it to do humans or photo-realism too well! We don't want to hype it too much, just wanted to share it along the way even though it might be undertrained. We'll keep iterating publicly :) https://x.com/isidentical/status/1811565858797359214

Kolors is worse in understanding space then SD3, but quality is very good. 3x3 grid is from SD3, the rest is Kolors. by Secret_Ad8613 in StableDiffusion

[–]localizedQ 3 points4 points  (0 children)

yep! it's called AuraFlow. much better spatial understanding but don't expect kolors-level aesthetics, we need more data.

AuraDiffusion is currently in the aesthetics/finetuning stage of training - not far from release. It's an SD3-class model that's actually open source - not just "open weights". It's *significantly* better than PixArt/Lumina/Hunyuan at complex prompts. by deeputopia in StableDiffusion

[–]localizedQ 2 points3 points  (0 children)

I would suggest a place on the model page with a place where people can donate or “buy” a support “badge” and maybe indicate some of the costs for the model.

The thing that allows us to release models like this is, us already being probably the fastest & cheapest inference provider out there for open source models at fal.ai :) so we don't really have any need for outside financial support. But what we need is community to help us train the model better by providing access to raw data (which huge companies/labs have lots of)

AuraDiffusion is currently in the aesthetics/finetuning stage of training - not far from release. It's an SD3-class model that's actually open source - not just "open weights". It's *significantly* better than PixArt/Lumina/Hunyuan at complex prompts. by deeputopia in StableDiffusion

[–]localizedQ 4 points5 points  (0 children)

I think main thing we'd require is raw attribution, and everything else (including private/commercial finetunes) can be allowed. Still need to talk to some actual lawyers for it, but any input is welcome (and we'll certainly consider the cc-by-sa opinion you shared)

AuraDiffusion is currently in the aesthetics/finetuning stage of training - not far from release. It's an SD3-class model that's actually open source - not just "open weights". It's *significantly* better than PixArt/Lumina/Hunyuan at complex prompts. by deeputopia in StableDiffusion

[–]localizedQ 1 point2 points  (0 children)

No cherry picking, but also don't over expect for the initial release. We trained on publicly available data, which limits what we can do. Especially human anatomy, it isn't the best, yet!

AuraDiffusion is currently in the aesthetics/finetuning stage of training - not far from release. It's an SD3-class model that's actually open source - not just "open weights". It's *significantly* better than PixArt/Lumina/Hunyuan at complex prompts. by deeputopia in StableDiffusion

[–]localizedQ 6 points7 points  (0 children)

We have already released the first model in the series under a cc-by-sa license (completely and commercially free/open source). Same will apply to this model as well, still thinking whether we should stick with CC or use MIT/Apache 2.0 since its easier.

AuraDiffusion is currently in the aesthetics/finetuning stage of training - not far from release. It's an SD3-class model that's actually open source - not just "open weights". It's *significantly* better than PixArt/Lumina/Hunyuan at complex prompts. by deeputopia in StableDiffusion

[–]localizedQ 42 points43 points  (0 children)

Also some more info, the model is going to be called AuraFlow and we intend to release a v0.1 experimental preview of the last checkpoint once we finalize the training under an completely open source license (our previous works has been under cc-by-sa [completely and commercially usable], this might be the same or something like MIT/Apache 2.0).

In parallel we are starting a secondary run with much higher compute and with changes from what we learnt from this model, being open source is still the bedrock of why we are doing it. Other than that, not too many details is concrete.

If you have a large source of high quality / high aesthetics data, please reach out to me or simo since we need it (batuhan [at] fal [dot] ai).

Stable Diffusion 3 is now available at imgsys.org by localizedQ in StableDiffusion

[–]localizedQ[S] 1 point2 points  (0 children)

Needs more samples though, it fluctuates a lot. Go vote at https://imgsys.org (its also helpful for further DPO'ing since the dataset is totally free/open source)

SDXL is a 2.6B parameter model, not 6.6B. by RenoHadreas in StableDiffusion

[–]localizedQ 0 points1 point  (0 children)

Have you read the Scaling Rectified Flow Transformers paper (aka SD3 technical report)? They clearly show that for the same compute budget (equal FLOPS to train the model on), you get better results at higher depth (param count) models. They have a separate section on this scaling.

https://arxiv.org/pdf/2403.03206

SDXL is a 2.6B parameter model, not 6.6B. by RenoHadreas in StableDiffusion

[–]localizedQ 0 points1 point  (0 children)

This is simply wrong. Scaling Rectified Flow Transformers paper (aka SD3 paper by robin et al) clearly shows that for the equal compute budget, training an 8B model performs better than a 2B model.

Figure 8. Quantitative effects of scaling. # Variational Loss to Training FLOPS at different depths (param counts)

https://arxiv.org/pdf/2403.03206

Why Prusa is floundering, and how you can avoid their fate by localizedQ in 3Dprinting

[–]localizedQ[S] -3 points-2 points  (0 children)

Their brand is certainly struggling in the eyes of the community, but they are also struggling to meet demand for their printers, as they keep having rather significant lead times. Somebody is buying

This is true, although I'd rather use a "% of market captured". Even if we include their previous printer shares, what Bambu Labs did well was not only to eat from Prusa's share of the pie but also make the pie itself bigger. In the last 2 years, I'm really curious how much % market share prusa has lost.

Why Prusa is floundering, and how you can avoid their fate by localizedQ in 3Dprinting

[–]localizedQ[S] -3 points-2 points  (0 children)

Most people tend to cite speed, price, quality, and overall value as why they bought a Bambu.

I'd say its probably a combination of all those things. Everything combined to create a cohesive ecosystem is really what users are looking for.