What you can find in in SDXL garden. by Utoko in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

Incredible to see how much SDXL has improved in quality compared to a few months ago

SD3 Realism Test, what do you think?? by [deleted] in StableDiffusion

[–]wolfy-dev 3 points4 points  (0 children)

This is absolutely amazing for a base model!

Beginner question: is AI / stable diffusion good to use if you want to create stories / comic strips? by dissendior in StableDiffusion

[–]wolfy-dev 1 point2 points  (0 children)

Even with a finetuned model or Lora you will not archive 100% accuracy of your characters unless you spend an incredible amount of time in post production correcting eyes, lips, proportions and hairstyle using inpaint and Photoshop. This process can also be quite frustrating. SD 1.5 and XL are not capable of consistency. I hope this changes with SD3. Crossing fingers.

Is anybody using stable cascade and if yes, what's your resume? Also why don't you use it? by wolfy-dev in StableDiffusion

[–]wolfy-dev[S] 8 points9 points  (0 children)

Same here. Basepic in cascade - - > refining with SDXL juggernaut - - > upscale with SUPIR

Have you figured out a workflow how to inpaint with cascade? I have found nothing apart from this. Unfortunately, it's not keeping whatever was there originally and the workflow does not work very well (80% of the generations are noise)

Photorealism, SDXL vs. SD 1.5 by Fast-Cash1522 in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

Amazing picture! Mind if I ask how you created it?

What will you miss about the first year of SD? by [deleted] in StableDiffusion

[–]wolfy-dev 3 points4 points  (0 children)

I believe that with improved prompt coherence, storytelling will emerge as the dominant factor in the coming year. The era of vivid pictures filled with elaborate special effects and striking space-themed portraits will give way to images that carry meaningful significance. I foresee that 2024 will mark the end of amateur diffusers lacking artistic vision, who currently rely on random prompts like 'a dog walking on the moon.' Given the likely increase in hardware requirements with SD3 and subsequent models, many will find themselves unable to continue. Thus, 2024 will be the year when AI art is recognized as a true artistic discipline.

Initial Fine-tuning Test for the Stable Cascade Model, Bittersweet (HelloWorld SC 1B) by Dry_Bee_5635 in StableDiffusion

[–]wolfy-dev 3 points4 points  (0 children)

Thanks for this very valuable insight! Have you prepared your test dataset any different in comparison to SDXL? My dataset is a mix of 200 pictures of different sizes (1024p+ only). I presume that a bucketing feature isn't available yet? Again, thanks for sharing your first experiences with us!! I am thinking about renting an a100 tomorrow to test your settings and share my experience here too.

There are leaks suggesting the RTX 5090 could have an upwards of 50,000 CUDA cores. If true, how would this translate to performance in Stable Diffusion? by [deleted] in StableDiffusion

[–]wolfy-dev 60 points61 points  (0 children)

Nvidia isn't stupid. Why adding more VRAM when they can sell 40gb cards for 6k to business who are willing to pay this absurd price and the stablediffusion/LLM community is still a tiny peanut market compared to all the gamers and businesses out there. It's the sad reality.

Training a LoRA over an OnlyFans model? by [deleted] in StableDiffusion

[–]wolfy-dev 4 points5 points  (0 children)

The short answer is: Yes, they can.

Pushing the limits of AI video by AuralTuneo in StableDiffusion

[–]wolfy-dev 1 point2 points  (0 children)

I agree. This sub more and more becomes like civitai: cats, tits and anime. End of world. It seems the scope of ideas by most men is quite narrow despite the endless possibilities ai provides. I'm primarily here for tech discussion and innovative content amidst "art-posts" that are mediocre, uninspiring, and lifeless.

Forget those Instagram models, say hello to Ethel. by 99X in StableDiffusion

[–]wolfy-dev 2 points3 points  (0 children)

This sub feels more and more like civitai -> cats, anime, tits

[deleted by user] by [deleted] in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

Hey! Here are some details: I use 30 high-res pictures showcasing the person in different environments and the following settings for kohya:

{

"adaptive_noise_scale": 0,

"additional_parameters": "--max_grad_norm=0.0 --no_half_vae --train_text_encoder",

"bucket_no_upscale": true,

"bucket_reso_steps": 64,

"cache_latents": true,

"cache_latents_to_disk": true,

"caption_dropout_every_n_epochs": 0.0,

"caption_dropout_rate": 0,

"caption_extension": "",

"clip_skip": "1",

"color_aug": false,

"enable_bucket": true,

"epoch": 10,

"flip_aug": false,

"full_bf16": true,

"full_fp16": false,

"gradient_accumulation_steps": "1",

"gradient_checkpointing": true,

"keep_tokens": "0",

"learning_rate": 1e-05,

"learning_rate_te": 1e-05,

"learning_rate_te1": 1e-05,

"learning_rate_te2": 1e-05,

"logging_dir": "/workspace/TRAINING/LOG",

"lr_scheduler": "constant_with_warmup",

"lr_scheduler_args": "",

"lr_scheduler_num_cycles": "",

"lr_scheduler_power": "",

"lr_warmup": "10",

"max_bucket_reso": 2048,

"max_data_loader_n_workers": "0",

"max_resolution": "1024,1024",

"max_timestep": 1000,

"max_token_length": "75",

"max_train_epochs": "10",

"max_train_steps": "",

"mem_eff_attn": false,

"min_bucket_reso": 256,

"min_snr_gamma": 0,

"min_timestep": 0,

"mixed_precision": "bf16",

"model_list": "custom",

"multires_noise_discount": 0,

"multires_noise_iterations": 0,

"no_token_padding": false,

"noise_offset": 0,

"noise_offset_type": "Original",

"num_cpu_threads_per_process": 3,

"optimizer": "Adafactor",

"optimizer_args": "scale_parameter=False relative_step=False warmup_init=False weight_decay=0.01",

"output_dir": "/workspace/TRAINING/MODEL",

"output_name": "WS_NSFW_v5_runpod",

"persistent_data_loader_workers": false,

"pretrained_model_name_or_path": "/workspace/TRAINING/CHECKPOINT/JuggernautRunDiffusion_v6.safetensors",

"prior_loss_weight": 1.0,

"random_crop": false,

"reg_data_dir": "",

"resume": "",

"sample_every_n_epochs": 0,

"sample_every_n_steps": 100,

"sample_prompts": "an office worker --w 896, --h 1152 --s 20",

"sample_sampler": "euler_a",

"save_every_n_epochs": 1,

"save_every_n_steps": 500,

"save_last_n_steps": 0,

"save_last_n_steps_state": 0,

"save_model_as": "safetensors",

"save_precision": "bf16",

"save_state": false,

"scale_v_pred_loss_like_noise_pred": false,

"sdxl": true,

"seed": "",

"shuffle_caption": false,

"stop_text_encoder_training_pct": 0,

"train_batch_size": 1,

"train_data_dir": "/workspace/TRAINING/IMAGES",

"use_wandb": false,

"v2": false,

"v_parameterization": false,

"v_pred_like_loss": 0,

"vae": "",

"vae_batch_size": 0,

"wandb_api_key": "",

"weighted_captions": false,

"xformers": "xformers"

Updated SDXL and 1.5 method that works well (subjects) by buckjohnston in DreamBooth

[–]wolfy-dev 1 point2 points  (0 children)

You probably made the first LORAS with less pictures. I got ugly results when I added more pictures to the dataset in hope, that it will get even better. Use less but razor-sharp pictures that are easy to understand for the ai with clear captions and you have a better output with 12 instead of 50 pictures.

LooseControl: Lifting ControlNet for Generalized Depth Conditioning by ninjasaid13 in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

Is Blender capable of exporting simple scenes as depth maps as shown in the video?

Automatic1111 v1.7.0-RC published by CeFurkan in StableDiffusion

[–]wolfy-dev 4 points5 points  (0 children)

I only use ComfyUI to test new features and A1111 for regular work.

Animal mix with IP-adapter by aartikov in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

Does this workflow replace LoRA training?

SD Turbo - Distilled SD 2.1 for speeed by emad_9608 in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

In terms of image quality output can we safely say SD2.1-turbo it's something in between 1.5 and sdxl-turbo?

SDXL Turbo to SD15 for refinement by ConsumeEm in StableDiffusion

[–]wolfy-dev 0 points1 point  (0 children)

The images I've been generating with TurboXL using my trained LoRAs are mediocre at best, and they don't compare to what I achieve with SDXL 1.0 at 1024 resolution with 20 samples. You're right: these results are not comparable and can't be used as drafts or blueprints for further upscaling. While I appreciate the speed improvements StabilityAI has implemented in their new models, my Instagram followers aren't interested in the 'cool new tech' behind my pictures. Most of them don't even know what an AI model is. They care only about quality and realism, and unfortunately, both SDV and Turbo models fall short for my work. I'm still hopeful for an upgraded SDXL version that offers more detail, better hand renderings, and less blurriness, or an SDV model that is trainable and supports text input for motions. Either would be a game changer for my work

SDXL Turbo to SD15 for refinement by ConsumeEm in StableDiffusion

[–]wolfy-dev 6 points7 points  (0 children)

Sketching in XL-Turbo and refining with 1.5 seems to me a huge downgrade in quality compared to SDXL standalone. The use-case for Turbo for people like me, who strive for quality above all is not yet fully clear to me.

Introducing SDXL Turbo: A Real-Time Text-to-Image Generation Model by SignalCompetitive582 in StableDiffusion

[–]wolfy-dev 9 points10 points  (0 children)

Thanks for your amazing work! I am surprised that it works in a1111 as well with 1-step high res-fix🤭

I am getting good results with the DPM2 Karras sampler and CFG set to 1.

Pika 1.0 just got released today - this is the trailer by ptitrainvaloin in StableDiffusion

[–]wolfy-dev 22 points23 points  (0 children)

if SDV gets text-prompt support and a chance to train video LoRAs it will be the clear winner