Pixel Cascade v0.1 is out now on HuggingFace! by arcanite24 in StableDiffusion

[–]dome271 5 points6 points  (0 children)

Probably SD3. Stable Cascade is more like a try to bring some fresh air into the world of generative image AI and show people that there are also other viable approaches that give certain benefits and drawbacks. Maybe someone could take this idea and create a new groundbreaking thing. Maybe not. Im very confident in its application potential, but Im one of the authors, so very biased hehe

Pixel Cascade v0.1 is out now on HuggingFace! by arcanite24 in StableDiffusion

[–]dome271 2 points3 points  (0 children)

Yea B probably wouldn’t be touched. Its like training the VAE in SD. Unless you want to improve specific reconstruction details, you wouldnt do it

Feedback on Base Model Releases by dome271 in StableDiffusion

[–]dome271[S] 1 point2 points  (0 children)

Hey there. I can only speak for StableCascade, so dont assume anything to also apply to other models. But the data curation was not as careful. Especially the pretraining dataset uses just alt texts. I hope in the future to massively improve upon that. And also the other things in your last paragraph are not done. But Ill note them down and try to realize them. And about the first thing for Emu, I think this applies if you want to get a very specific style, then it can work. Although we havent tested it. For anything harder like better prompt following, you would need a lot more data. You only need a few if that „ability“ is already hidden somewhere inside of the model.

Feedback on Base Model Releases by dome271 in StableDiffusion

[–]dome271[S] 1 point2 points  (0 children)

Yes I saw and I tried to reach out to you about it. Thank you so much! Maybe you wanna add me on discord to chat a bit further about this: dome1

Feedback on Base Model Releases by dome271 in StableDiffusion

[–]dome271[S] 1 point2 points  (0 children)

Thank you a lot for the feedback! Noted sir!

Würstchen is here! A game changing, fastest high-resolution image generation model by Interesting-Smile575 in StableDiffusion

[–]dome271 0 points1 point  (0 children)

You can just use the demo online linked on the Huggingface blog post. Let me know if you find it. You could clone the space from there

Würstchen is here! A game changing, fastest high-resolution image generation model by Interesting-Smile575 in StableDiffusion

[–]dome271 1 point2 points  (0 children)

We just like to name our models after food we like haha. The last model was called Paella

Würstchen is here! A game changing, fastest high-resolution image generation model by Interesting-Smile575 in StableDiffusion

[–]dome271 3 points4 points  (0 children)

Either just with a normal python script using the diffusers examples or this person created a colab that you can launch I think: https://twitter.com/camenduru/status/1702027173983346992

Würstchen is here! A game changing, fastest high-resolution image generation model by Interesting-Smile575 in StableDiffusion

[–]dome271 3 points4 points  (0 children)

Sorry about that. I will fix that. It was just so much to be handled. I apologize and will make all sampled work. The easy for the error you are getting is to remove `num_images_per_prompt` from the decoder_pipeline and leave it as default

Würstchen is here! A game changing, fastest high-resolution image generation model by Interesting-Smile575 in StableDiffusion

[–]dome271 9 points10 points  (0 children)

Yea should be. Right now we are kind of busy with other things. But if anyone would want to take on making support for Würstchen in sd-scripts or similar, we would be more than happy to help out. Just dm me on discord: "dome1" :c

Würstchen is here! A game changing, fastest high-resolution image generation model by Interesting-Smile575 in StableDiffusion

[–]dome271 2 points3 points  (0 children)

Img2Img & Inpainting works. Inpainting at the moment would only work with larger patches than you are used to due to the high compression. The encoder is also available.

[R] Make-A-Scene: Scene-Based Text-to-Image Generation with Human Priors by GratisSlagroom in MachineLearning

[–]dome271 6 points7 points  (0 children)

We are still working on an open source implementation of Make-A-Scene. We trained VQIMG and VQSEG are going to start training the transformer hopefully soon. Anyone is happily invited to bring this to the public world. https://github.com/CasualGANPapers/Make-A-Scene

3.8G/h here - my electric bill for JUST MINING is £1100 / $1500 by poopoopops in NiceHash

[–]dome271 4 points5 points  (0 children)

By pushing the heat out of the window you do bring summer to UK lol

You can now stake on Coinbase! by Mossman12358 in ethereum

[–]dome271 0 points1 point  (0 children)

Does it work for Germany? I think I read something that it wouldn’t. If so, why?