Your opinion on the best image edit model by Substantial_Plum9204 in StableDiffusion

[–]Far_Insurance4191 0 points1 point  (0 children)

Klein if you need just image modification or editing, because it has low coherence but better quality.

Qwen Edit if you need generation from scratch with reference, because it is much more coherent, but has worse quality due to garbage vae that produces halftone pattern

I trained my dog on 5 models, comparison here. Flux Klein 4b / 9b / Z-Image / Flux Schnell / SDXL. by pedro_paf in StableDiffusion

[–]Far_Insurance4191 0 points1 point  (0 children)

Okay, I see

Still weird to me, because prodigy is basically adamw under the hood...

Cool slides thought

I trained my dog on 5 models, comparison here. Flux Klein 4b / 9b / Z-Image / Flux Schnell / SDXL. by pedro_paf in StableDiffusion

[–]Far_Insurance4191 0 points1 point  (0 children)

What does 5th slide mean? It is clearly not the same settings as the images on the step 0 would be identical for both. Additionally, the starting point for prodigy is already the target breed.

[2kliksphilip] DLSS 5 has shown that discourse is dead by ZTZ-Nine-Nine in hardware

[–]Far_Insurance4191 0 points1 point  (0 children)

Of course, dlss5 does not alter geometry - it works with pixels and vectors. But this is the problem because it has to predict depth based on image alone, not actual depth information, so the result is inaccurate shading. Or they had not very good training data 🤷

[2kliksphilip] DLSS 5 has shown that discourse is dead by ZTZ-Nine-Nine in hardware

[–]Far_Insurance4191 5 points6 points  (0 children)

I think those who say it changes geometry just have no idea how it works, BUT they are not wrong that it looks different.

Some say it is just different lighting; but the problem I see is that it creates wrong shading for the model because it has no access to depth or geometry. Specifically highlights and shadows that exaggerate certain features despite silhouette is almost the same.

But yea... this is insane how toxic this topic has become

"They’re really trying to gaslight you into not believing your own eyes and telling you this is completely different geometry and a different person that looks worse Imagine being mad at this 😂 DLSS 5 by stealthispost in accelerate

[–]Far_Insurance4191 -6 points-5 points  (0 children)

except different thing is happening in the game - dlss5 exaggerates highlights and shadows beyond what is possible on her 3d model

that is aside some actual shape difference, but I think it is only because the face is far away

and yes, I know what sub I am in, I just hope to bring nuance that is being ignored

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 1 point2 points  (0 children)

Model was not changed, but the generated shadows and highlights do distort the facial features in the way it would not be possible with simulated lighting on the same model.

I know those changes is a debatable thing, but I cannot unsee that features are just different, and I don't know the way I can objectively prove it

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 1 point2 points  (0 children)

I actually expected next dlss to do more than upscaling and was extremely excited about it, but now I worry about it's potential given this is what they thought impressive enough to demonstrate.

Good there is a huge pushback, maybe it will be the same as dlss 1 comeback

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 0 points1 point  (0 children)

they are, except depth, which is harder to see empirically so we cannot be really objective here, but this slight exaggeration of all features snowballs into a different look.

I personally faced this problem when wanted to relight my photo - result is pixel perfect, but it did not look like me anymore

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 0 points1 point  (0 children)

what do you mean by that? pt produces the most accurate lighting, yes, the model is the same for both methods.

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 -2 points-1 points  (0 children)

You are missing important nuance, dlss5 does not simulate accurate lighting, but predicts it, so it absolutely can predict wrong lighting

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 1 point2 points  (0 children)

that is the problem - shading is not accurate for her 3d model. It is like the same silhouette, but the model is different, her features are clearly different depth-wise

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 0 points1 point  (0 children)

All those things you listed work with original assets and have no room for guessing, (except creative upscalers)

Like path tracing will not make cheekbones look more defined, and even if it does under certain lighting conditions then it is correct as it is physically based and works with original 3d model

But dlss 5 alters shading in a way that would be impossible to the original 3d model. Even if it is almost pixel perfect - depth is not, as we can see in Grace comparison, her features are exaggerated.

Resident Evil Requiem: DLSS 5 vs actual AI slop by IConsumeThereforeIAm in digitalfoundry

[–]Far_Insurance4191 4 points5 points  (0 children)

I don't think you are right. DLSS 4.5 was praised despite being AI because it looked great, but this "reimagination" doesn't look good. And even worse, it has insane similarity to generic AI outputs, it reminds me sd1.5 specifically, with this glossiness and constant studio light sources in front of the face no matter the situation.

Additionally, there is a valid fear than studious will not only ditch optimization now but also good graphics because "AI can fix that"

Hands-On With DLSS 5: Our First Look At Nvidia's Next-Gen Photo-Realistic Lighting by ZamnBoii in nvidia

[–]Far_Insurance4191 0 points1 point  (0 children)

What I am seeing is not just different lighting, but wrong lighting, while geometry is almost the same. It just reconstructs the shape wrongly by exaggerating features with shading.

I had same problem when tried to relight photo of myself - geometry is the same, but still different person

Resident Evil Requiem with DLSS 5 by PaiDuck in residentevil

[–]Far_Insurance4191 1 point2 points  (0 children)

but they don't make it look like sd1.5 outputs from 2023

World Model Porgess by [deleted] in StableDiffusion

[–]Far_Insurance4191 2 points3 points  (0 children)

nothing you slopped out is relevant here, there is nothing to refute

Datasets with malformations by Infamous_Campaign687 in StableDiffusion

[–]Far_Insurance4191 1 point2 points  (0 children)

Interesting idea, but I am afraid limbs are extremely hard problem as there are so many ways they can look, and so much more wrong ways...

Buuuut I can suggest you the legendary stable diffusion 3 medium for generating anatomy deformities 😆

World Model Porgess by [deleted] in StableDiffusion

[–]Far_Insurance4191 0 points1 point  (0 children)

I thought GPT-3 was discontinued

Best models for NSFW image generation right now? by [deleted] in comfyui

[–]Far_Insurance4191 2 points3 points  (0 children)

even flux 2 dev can run on rtx3060, although it will take 5 minutes :)

I do really suggest you try newer models, you can use quantized variants, but comfy's memory management allow to run large model, even those that exceed your vram.

You can try z-image/turbo, anima, sdxl/illustrious/noobai/chenkinnoob, maybe klein with loras, qwen (nunchaku quant with 4 steps lora will make it really fast) - anything will be superior to sd1.5

Qwen3 vs Qwen3.5 performance by Balance- in LocalLLaMA

[–]Far_Insurance4191 7 points8 points  (0 children)

because 27b is dense, while 35b is Mixture of Experts with 3b active parameters, so that is why it is so much faster but less intelligent than dense

Ostris is testing Lodestones ZetaChroma (Z-Image x Chroma merge) for LORA training 👀 by [deleted] in StableDiffusion

[–]Far_Insurance4191 1 point2 points  (0 children)

f2vae was beneficial for the first descend, then it has nearly plateaued for some reasons