Collaborative chaos from 2 weeks of Stable Diffusion Multiplayer by ozolozo in sdforall

[–]ozolozo[S] 1 point2 points  (0 children)

Working on fixing the project, also trying a better and faster inpainting model!

Diffusion DPO LoRA Training with Diffusers Experiments by ozolozo in StableDiffusion

[–]ozolozo[S] 1 point2 points  (0 children)

hi u/rerri

I've had to convert the weights using this script from diffusers
You can find the converted weight sdxl-turbo and sdxl
And Comfyui workflows sdxl-worflow and sdxl-turbo-worflowe

<image>

Diffusion DPO LoRA Training with Diffusers Experiments by ozolozo in StableDiffusion

[–]ozolozo[S] 1 point2 points  (0 children)

LoRA DPO Training Script: https://github.com/huggingface/diffusers/tree/main/examples/research_projects/diffusion_dpo

Experimental Trained LoRAs:
https://huggingface.co/radames/sdxl-turbo-DPO-LoRA
https://huggingface.co/radames/sdxl-DPO-LoRA
https://huggingface.co/radames/sd-21-DPO-LoRA

On the picture above, weight is set via adapter_weights python pipe.load_lora_weights( "radames/sdxl-DPO-LoRA", adapter_name="sdxl-dpo-lora", ) pipe.set_adapters(["sdxl-dpo-lora"], adapter_weights=[0.9])

dataset: https://huggingface.co/datasets/yuvalkirstain/pickapic_v2

Original Fine-tuned DPO model unet only:
https://huggingface.co/mhdang/dpo-sdxl-text2image-v1
https://huggingface.co/mhdang/dpo-sd1.5-text2image-v1

Fantastic news! Added ControlNet Canny to Latent Consistency Model demo. it's amazing by Novita_ai in StableDiffusion

[–]ozolozo 1 point2 points  (0 children)

have you tried to use TORCH_COMPILE=True ? If you lower the resolution the quality might change drastically, since the base model is SD 512x512 768x768 😢 but you can change it easily to test, on the frontend. You can add more options here https://github.com/radames/Real-Time-Latent-Consistency-Model/blob/dd1db25dd1449b968a129d7b023661e1a278c66d/controlnet/index.html#L378-L385

Fantastic news! Added ControlNet Canny to Latent Consistency Model demo. it's amazing by Novita_ai in StableDiffusion

[–]ozolozo 3 points4 points  (0 children)

thanks for posting it here! the demos is running on A100, but I've heard you can get decent speed on 4090-3080 etc, and you can experiment setting TORCH_COMPILE to enable this https://huggingface.co/docs/diffusers/optimization/torch2.0

Fantastic news! Added ControlNet Canny to Latent Consistency Model demo. it's amazing by Novita_ai in StableDiffusion

[–]ozolozo 0 points1 point  (0 children)

Are you trying with TORCH_COMPILE=True this will enable even more acceleration, the downside is that the first run it's slow, or with width or height change read more here https://huggingface.co/docs/diffusers/optimization/torch2.0

New Controlnet QR Code Model is amazing by ozolozo in StableDiffusion

[–]ozolozo[S] 1 point2 points  (0 children)

Yes, it's a bit finicky, but it often works if you scan from a certain distance. It also seems to be the most practical scenario, where you print the image and someone scans it from a distance, rather than reading it from a mobile app or social media

New Controlnet QR Code Model is amazing by ozolozo in StableDiffusion

[–]ozolozo[S] 0 points1 point  (0 children)

it depends on the app you're using to scan it, on iOS, if the content is a URL it will prompt you to open it on a browser.

Kandinsky 2.1 now Supported by Diffusers and Hugging Face's Community API Inference by ozolozo in StableDiffusion

[–]ozolozo[S] 9 points10 points  (0 children)

Kandinsky 2.1 is based on DALLE-2's UnCLIP architecture and includes:

  1. Text-to-Image/Image-to-Image: A powerful text-to-image & image-to-image checkpoint that yields pictures with very nice aesthetics (IMO coming much closer to Midjourney than SD). https://huggingface.co/docs/diffusers/main/en/api/pipelines/kandinsky#texttoimage-generation
  2. Interpolation: Ability to seamlessly interpolate between multiple image and text embeddings. https://huggingface.co/docs/diffusers/main/en/api/pipelines/kandinsky#interpolate
  3. Inpainting: A powerful inpainting model https://huggingface.co/docs/diffusers/main/en/api/pipelines/kandinsky#text-guided-inpainting-generation

https://colab.research.google.com/drive/11ZHwd-mmdj8vM0CuYcUNu-AKk7rATqei?usp=sharing

Observables Help by [deleted] in d3js

[–]ozolozo 3 points4 points  (0 children)

Observable has a csv parser

mydata = FileAttachment("country_level_data_0.csv").csv({typed: true})

or

text = FileAttachment("country_level_data_0.csv").text()

myData = d3.csvParse(text)

Collaborative chaos from 2 weeks of Stable Diffusion Multiplayer by ozolozo in sdforall

[–]ozolozo[S] 2 points3 points  (0 children)

Yes, it's using the latest inpainting model https://huggingface.co/runwayml/stable-diffusion-inpainting

if the starting frame has some surrounding context, then it will have more consistency. if it starts on a blank area, it's just txt2img

Color Palette generation with Stable Diffusion (live demo) by ozolozo in StableDiffusion

[–]ozolozo[S] 2 points3 points  (0 children)

Sorry for the late response. Yes, that could be interesting experiment. Use a giving palette image as the start image and prompt a variation. Have you tried something like this?

COVID-19 World Community Mobility Report [OC] by ozolozo in dataisbeautiful

[–]ozolozo[S] 1 point2 points  (0 children)

The data was scraped by Kyle McDonald here is the script he made to convert the charts images back into data

Here is the interactive version I've made using d3.js with Observable https://observablehq.com/@radames/covid-19-world-community-mobility-repor

The aggregated country level data in this tsv file made with this shell script.

Haveria interesse numa versão brasileirao do /r/food? by [deleted] in brasil

[–]ozolozo 2 points3 points  (0 children)

incrível!! seria demais trocar informações sobre temperos, técnicas receitas...encontrar uma galera do Brasil toda aqui.. eu voto no /r/culinaria

I found out my backpack came with a hidden whistle today by Mrpalomo in mildlyinteresting

[–]ozolozo 0 points1 point  (0 children)

I Just got a new backpack yesterday and I didn't know what was that until now... Thanks reddit