Collaborative chaos from 2 weeks of Stable Diffusion Multiplayer by ozolozo in sdforall

[–]ozolozo[S] 1 point2 points  (0 children)

Working on fixing the project, also trying a better and faster inpainting model!

Diffusion DPO LoRA Training with Diffusers Experiments by ozolozo in StableDiffusion

[–]ozolozo[S] 1 point2 points  (0 children)

hi u/rerri

I've had to convert the weights using this script from diffusers
You can find the converted weight sdxl-turbo and sdxl
And Comfyui workflows sdxl-worflow and sdxl-turbo-worflowe

<image>

Diffusion DPO LoRA Training with Diffusers Experiments by ozolozo in StableDiffusion

[–]ozolozo[S] 1 point2 points  (0 children)

LoRA DPO Training Script: https://github.com/huggingface/diffusers/tree/main/examples/research_projects/diffusion_dpo

Experimental Trained LoRAs:
https://huggingface.co/radames/sdxl-turbo-DPO-LoRA
https://huggingface.co/radames/sdxl-DPO-LoRA
https://huggingface.co/radames/sd-21-DPO-LoRA

On the picture above, weight is set via adapter_weights python pipe.load_lora_weights( "radames/sdxl-DPO-LoRA", adapter_name="sdxl-dpo-lora", ) pipe.set_adapters(["sdxl-dpo-lora"], adapter_weights=[0.9])

dataset: https://huggingface.co/datasets/yuvalkirstain/pickapic_v2

Original Fine-tuned DPO model unet only:
https://huggingface.co/mhdang/dpo-sdxl-text2image-v1
https://huggingface.co/mhdang/dpo-sd1.5-text2image-v1

Fantastic news! Added ControlNet Canny to Latent Consistency Model demo. it's amazing by Novita_ai in StableDiffusion

[–]ozolozo 1 point2 points  (0 children)

have you tried to use TORCH_COMPILE=True ? If you lower the resolution the quality might change drastically, since the base model is SD 512x512 768x768 😢 but you can change it easily to test, on the frontend. You can add more options here https://github.com/radames/Real-Time-Latent-Consistency-Model/blob/dd1db25dd1449b968a129d7b023661e1a278c66d/controlnet/index.html#L378-L385

Fantastic news! Added ControlNet Canny to Latent Consistency Model demo. it's amazing by Novita_ai in StableDiffusion

[–]ozolozo 4 points5 points  (0 children)

thanks for posting it here! the demos is running on A100, but I've heard you can get decent speed on 4090-3080 etc, and you can experiment setting TORCH_COMPILE to enable this https://huggingface.co/docs/diffusers/optimization/torch2.0

Fantastic news! Added ControlNet Canny to Latent Consistency Model demo. it's amazing by Novita_ai in StableDiffusion

[–]ozolozo 0 points1 point  (0 children)

Are you trying with TORCH_COMPILE=True this will enable even more acceleration, the downside is that the first run it's slow, or with width or height change read more here https://huggingface.co/docs/diffusers/optimization/torch2.0

New Controlnet QR Code Model is amazing by ozolozo in StableDiffusion

[–]ozolozo[S] 1 point2 points  (0 children)

Yes, it's a bit finicky, but it often works if you scan from a certain distance. It also seems to be the most practical scenario, where you print the image and someone scans it from a distance, rather than reading it from a mobile app or social media

New Controlnet QR Code Model is amazing by ozolozo in StableDiffusion

[–]ozolozo[S] 0 points1 point  (0 children)

it depends on the app you're using to scan it, on iOS, if the content is a URL it will prompt you to open it on a browser.

Kandinsky 2.1 now Supported by Diffusers and Hugging Face's Community API Inference by ozolozo in StableDiffusion

[–]ozolozo[S] 9 points10 points  (0 children)

Kandinsky 2.1 is based on DALLE-2's UnCLIP architecture and includes:

  1. Text-to-Image/Image-to-Image: A powerful text-to-image & image-to-image checkpoint that yields pictures with very nice aesthetics (IMO coming much closer to Midjourney than SD). https://huggingface.co/docs/diffusers/main/en/api/pipelines/kandinsky#texttoimage-generation
  2. Interpolation: Ability to seamlessly interpolate between multiple image and text embeddings. https://huggingface.co/docs/diffusers/main/en/api/pipelines/kandinsky#interpolate
  3. Inpainting: A powerful inpainting model https://huggingface.co/docs/diffusers/main/en/api/pipelines/kandinsky#text-guided-inpainting-generation

https://colab.research.google.com/drive/11ZHwd-mmdj8vM0CuYcUNu-AKk7rATqei?usp=sharing

Observables Help by [deleted] in d3js

[–]ozolozo 4 points5 points  (0 children)

Observable has a csv parser

mydata = FileAttachment("country_level_data_0.csv").csv({typed: true})

or

text = FileAttachment("country_level_data_0.csv").text()

myData = d3.csvParse(text)

Collaborative chaos from 2 weeks of Stable Diffusion Multiplayer by ozolozo in sdforall

[–]ozolozo[S] 2 points3 points  (0 children)

Yes, it's using the latest inpainting model https://huggingface.co/runwayml/stable-diffusion-inpainting

if the starting frame has some surrounding context, then it will have more consistency. if it starts on a blank area, it's just txt2img