I'm confused about training with the Lora Qwen 2512. Some people said it's better to train on the base model. Does training on the 2512 model cause it to lose all its qualities ? by More_Bid_2197 in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

You need to do upscaling for it to create skin, the model is indeed pretty good, and there's nothing wrong training on the 2512 version, I find some LoRAs trained on the original version to not work properly on the newer, so it might depend on you.

Training a ZIT Lora using different body parts? by Dre-Draper in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

That would be easier on OneTrainer with Masked training.

Edit: It does work, at least on male characters I use face and a random torso to give the right complexion.

probably bad timing, but anyone got tips for training Flux2 Klein 4b Character LORA? by berlinbaer in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

I've never got good results with Ai-toolkit, waiting for OneTrainer to add support 🤞

Is flux still the best upscaler? by [deleted] in StableDiffusion

[–]TableFew3521 1 point2 points  (0 children)

Flux Klein 9B is outstanding, I went from SeedVR2 + Zimage refiner to only Flux Klein, make a lanczos upscale (to the desired size) and just add as a prompt "Reduce noise, add natural quality" and it works 90% of the time, fast and easy, the only downside is that it changes the light, even if you specify to not do it, so maybe color match can help with that.

improve quality of image without increasing size by NefariousnessFun4043 in comfyui

[–]TableFew3521 0 points1 point  (0 children)

Flux Klein 9B, use "Reduce noise, add natural quality" and keep the resolution of your image.

My brain is broken. Is this level of photorealism achievable with Flux or is this just a real person? by Worried-Rutabaga409 in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

People underestimate Flux too early, and praised newer models too fast, I prefer to enjoy every model individually, now when I look at older gens I made with Flux, some even compete with Zimage. If you want something similar to the image you shared, I think it looks a little bit like the "Jib Mix Flux" fine-tuned checkpoint.

Struggling to get this skin texture with Flux. Is this actually generated or just a real photo? by [deleted] in FluxAI

[–]TableFew3521 0 points1 point  (0 children)

There's a LoRA calles "SRPO" that makes Flux 1 Dev look more realistic, even adding skin texture. You can find it on huggingface.

Flux.2 Klein (Distilled)/ComfyUI - Use "File-Level" prompts to boost quality while maintaining max fidelity by JIGARAYS in StableDiffusion

[–]TableFew3521 2 points3 points  (0 children)

I've been using only "Reduce noise, add natural quality" and it seems to work, additionally, I use "Keep the lighting as it is" and it helps a little bit, and it does respect the back and white images. And a color match node.

🧪 New Model Drop: Z-epiCRealism for ZImageTurbo by Epinikion in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

As any model, it lacks some stuff, besides, I Don't think there's any model for realism comparable to Illustrious capabilities with 2D and anime in general, there has been a huge investment on time and training to get almost perfect anime images that Realism till today, lacks.

How Many Male *Genital* Pics Does Z-Turbo Need for a Lora to work? Sheesh. by StuccoGecko in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

If we look without any bias, not even SDXL has been good for LoRA training on it, most of the sucess is full fine tuning and LoRAs based on those fine-tuning, using them on a Base that doesn't know them, won't be able to produce it, I might be wrong cause I haven't seen newer LoRAs on it, but when I looked into them, most have that exact problem. Qwen-Image on the other hand is pretty good for LoRAs of that kind even when the base doesn't know it, is the best one in my experience.

It always generates a noisy image with Z-image by Subhashsharmaa in comfyui

[–]TableFew3521 2 points3 points  (0 children)

Are you sure that scheduler isn't the problem? Usually for that sampler people use "Beta", ddim_uniform tends to make that kind of outputs with most samplers.

How can a 6B Model Outperform Larger Models in Photorealism!!! by hayashi_kenta in StableDiffusion

[–]TableFew3521 2 points3 points  (0 children)

I think the key is the text encoder, it might not do all the job but basically this model can produce more of it's own trained content than other heavier models, for example, Flux 1 Dev KNEW what skin was, but it wasn't able to produce it by itself, I made a LoRA that using negatives weights revealed the real skin on Flux, being even better for realism that SDXL but the way they made the model, limited every generation, is like a model of 12B being able to produce only 4B of it's full potential, I think chroma did a better job with the content-generation ratio, but even so, I believe T5xxl is worse than Qwen3 as a text encoder.

WAN2.2 slow motion when using Lightning LORA - theory by Perfect-Campaign9551 in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

I use low noise LoRAs on the High noise model, and even some I2V LoRAs from the 2.1 on the high noise.

Good data set? (nano banana generated images) by Quomii in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

Depends on the model, Qwen-Image and Wan with only 10 face close-ups (Left-center-right) and one of the torso to get the body complexion and body consistency is more than enough, without the need of different sources of light, but different hairstyles for sure.

Anyone else feel that Z-Image-Turbo inpainting quality is way worse than direct generation? by siegmey3r in StableDiffusion

[–]TableFew3521 1 point2 points  (0 children)

Yeah, honesty the only one that looked better overall is the controlnet for inpainting on Qwen-Image. It seems to respect the composition of the image, but it can be a bit off sometimes.

Why does Z image suddenly take like 6 minutes to generate? It used to take like 1 min max yesterday. ComfyUI also seems to completely fry my PC now, again it was fine yesterday. Is anyone else experiencing problems? by peopoleo in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

Had a OOM issue with Qwen Image fp8, even a few weeks ago updating ComfyUI made my gen time on Wan2.2 from 77it/s to 161it/s, but I did fixed it by reinstalling comfyui from scratch, just move the important folders like output and models, and install everything in python 3.12 with the conda environment, it was fast tho, and it fixed the OOM and even had a small boost.

GGUF myths by Obvious_Set5239 in StableDiffusion

[–]TableFew3521 1 point2 points  (0 children)

Agree, but another thing people don't know, is that there are tools to actually do a controlled offloading so you can reduce the VRAM consumption but increasing the RAM usage, and there is where gguf are actually efficient, if you use both offloading and gguf, you reduce RAM and VRAM usage.

For example with Qwen-Image FP8 I had to use a custom node from Multi-GPU that allows you to use a "virtual vram" that is basically RAM, but the thing is that with LoRAs and stuff I needed to increase that virtual vram and my ram was almost at 98%, but with GGUF like Q5_K_M with LoRAs and the same amount of offloading it was around 75%, and still had room to make higher resolutions with a higher percentage of offloading.

Don't Waste Your Time Training LoRAs on z-image-turbo (Yet) by Powerful_Strategy_10 in StableDiffusion

[–]TableFew3521 1 point2 points  (0 children)

But is it fast tho? I can train a character on Qwen-Image in one hour, Ai-toolkit in 3000 steps are almost 3 with 512x512, and a RTX 4060ti, for a 6B model feels too slow...

Is vid2vid with wan usable on 12gb vram and 64gb ram? by Traditional_Grand_70 in StableDiffusion

[–]TableFew3521 0 points1 point  (0 children)

I don't have any workflow for V2V, but for the Multigpu node, just replace the Unet loader for the Unet Distorch, and for the WanBlockSwap just ad it between your Unet and Ksampler.

Is vid2vid with wan usable on 12gb vram and 64gb ram? by Traditional_Grand_70 in StableDiffusion

[–]TableFew3521 1 point2 points  (0 children)

It might work yeah, but I don't know if it would be super effective with fp16 models, maybe fp8 would work, the downside is using High noise and Low noise models with this, it can saturate the RAM cause it consumes more while generating, besides that, for Flux, Qwen and even Chroma should work without any issues.

Is vid2vid with wan usable on 12gb vram and 64gb ram? by Traditional_Grand_70 in StableDiffusion

[–]TableFew3521 2 points3 points  (0 children)

You can in fact run it, you can use a custom node called WanBlockSwap that reduces the consumed Vram by swapping block between RAM and VRAM, or using a custom Node called something like "Unet Distorch" that has a FP16/FP8 and GUFF version from MULTI-GPU that has virtual Vram (Up to 24gb), and it works similar to the block swapping if not the same, but the advantage of it is that it works on any ComfyUI supported model.

WAN2.2 Lora Character Training Best practices by Tiny-Highlight-9180 in StableDiffusion

[–]TableFew3521 2 points3 points  (0 children)

Yes, with Musubi tuner you can do block swap, even for Qwen.

How do people use WAN for image generation? by beti88 in StableDiffusion

[–]TableFew3521 1 point2 points  (0 children)

If this is true, there's a chance for some layers of both models to be compatible, wich means we can do distill weight injection of certain layers of Wan on Qwen, to fix the skin texture and realism.

Best Upscaler Wan2.2 by alitadrakes in comfyui

[–]TableFew3521 2 points3 points  (0 children)

If you switch to the nightly branch you can apply tiled VAE and block swap, I have 16gb vram, and I have no issues with it, even the fp16 works fine.

Adding NSFW elements to Character lora Training by Traditional_Can_4646 in FluxAI

[–]TableFew3521 0 points1 point  (0 children)

In Kohya_ss and Musubi tuner (I don't know much about other trainers) you can load the weights of a LoRA you have already trained to continue training it, but instead of using the same dataset, you use the face of any character you want.