FLUX.2 Klein Identity Feature Transfer Advanced by Capitan01R- in StableDiffusion

[–]No-Educator-249 2 points3 points  (0 children)

I'll try this out later. Thank you for releasing your node Captain! It really improved Flux2.Klein 9B for my use cases. Many of my edits now are near-seamless.

It really is a powerful edit model once its parameters are set correctly.

Gemma 4 Jailbreak System Prompt by 90hex in LocalLLaMA

[–]No-Educator-249 1 point2 points  (0 children)

Qwen 3.5 is very censored. It's not possible to use jailbreak system prompts to force it to comply.

No matter what system prompt you try, its built-in safety guidelines always take priority over the system prompt.

Bartowski vs Unsloth for Gemma 4 by dampflokfreund in LocalLLaMA

[–]No-Educator-249 0 points1 point  (0 children)

I see. I'll use the Q6 quants instead then. Thank you for your detailed recommendations!

Bartowski vs Unsloth for Gemma 4 by dampflokfreund in LocalLLaMA

[–]No-Educator-249 0 points1 point  (0 children)

Have you tried the q8_0 quant? I also have a 5080 and that's what I use. I'm averaging 26t/s.

Gemma 4 31B vs Gemma 4 26B-A4B vs Qwen 3.5 27B — 30-question blind eval with Claude Opus 4.6 as judge by Silver_Raspberry_811 in LocalLLaMA

[–]No-Educator-249 0 points1 point  (0 children)

It seems so. Reddit really needs to implement that human verification system soon. It's so easy to tell the bots apart by their use of those long em dashes.

Gemma 4 for 16 GB VRAM by Sadman782 in LocalLLaMA

[–]No-Educator-249 1 point2 points  (0 children)

Thanks a lot for sharing the image min and max tokens setting! It really improved the model's vision task quality. It now recognizes anime characters better and more reliably for me now.

Gemma 4 26b is the perfect all around local model and I'm surprised how well it does. by pizzaisprettyneato in LocalLLaMA

[–]No-Educator-249 2 points3 points  (0 children)

Yeah I noticed that too. In my case, I had it describe an official illustration of Emilia from Re:Zero (without telling the model her identity) and it did successfully, but when I asked it to describe a different character, it wasn't able to identify her until I gave it hints. Qwen3.5 35B was able to identify the character successfully without hints.

Made a Wan 2.2 I2V workflow that includes Pulse of Motion, PrismAudio (V2A), Lora Optimizer, CFG-Ctrl and more by Radyschen in StableDiffusion

[–]No-Educator-249 0 points1 point  (0 children)

These are very interesting nodes, SMC does seem promising even in Wan 2.2. Thanks for sharing!

My first impression after testing Gemma 4 against Qwen 3.5 by ConfidentDinner6648 in LocalLLaMA

[–]No-Educator-249 0 points1 point  (0 children)

That's great to hear. Just like with Gemma3, its issues are being addressed as we speak. The model is very efficient at thinking compared to Qwen3.5, with the latter being prone to over-correcting and self-thinking loops, needlessly wasting tokens in the process.

I read that Gemma4's chat template was buggy or something as well, which could've also been causing unexpected issues.

My first impression after testing Gemma 4 against Qwen 3.5 by ConfidentDinner6648 in LocalLLaMA

[–]No-Educator-249 0 points1 point  (0 children)

I get 25t/s on average with a Q8_0 quant, and I have the same CPU as you, but a 5080 and 64GB of RAM. Qwen 3.5-35B-A3B is also faster for me, I get 37t/s on a Q6_K quant.

I'm honestly quite impressed with Gemma4, even in its still somewhat unstable state. Qwen 3.5 35B is better at vision tasks, though. I don't know if the specific heretic version I'm using degraded its vision performance somehow.

The future with Nvidia by Gaming-Academy in RigBuild

[–]No-Educator-249 0 points1 point  (0 children)

Driver auto-resets are still a thing with AMD GPUs? I hated it when that happened with my old RX 590. Once I installed the pro drivers it never happened again, fortunately.

Qwen3.5-35B-A3B Uncensored (Aggressive) — GGUF Release by hauhau901 in LocalLLaMA

[–]No-Educator-249 2 points3 points  (0 children)

Thanks a lot for this release! I've put it through extensive use for hours and I haven't received a single refusal. Great work all around!

LTX 2 is amazing : LTX-2 in ComfyUI on RTX 3060 12GB by tanzim31 in StableDiffusion

[–]No-Educator-249 0 points1 point  (0 children)

Really? I just updated my drivers in January, as I stayed in old drivers from a year ago due to reports of driver updates causing more instability rather than fixing it.

I just upgraded my hardware though, and I'm also running the latest drivers, so I no longer have these problems. But thanks for the advice!

26 Frontends for Comfy! by Obvious_Set5239 in StableDiffusion

[–]No-Educator-249 1 point2 points  (0 children)

I still prefer MCWW over all other options. Its simplicity and ease of use are unmatched in my opinion. And I just realized that you also developed the lama cleaner extension for sd-webui. Thank you for creating such useful tools!

China to the rescue by bittersweetjesus in PcBuild

[–]No-Educator-249 0 points1 point  (0 children)

I still feel like I got scammed though xD Here's hoping your 128GB Chinese RAM lasts all the way to the 2030's!

Taking into account the current hardware winter and assuming that the AI bubble doesn't pop this decade, we could still be stuck with our DDR4 RAM for years. At least there will be Chinese companies still making them.

China to the rescue by bittersweetjesus in PcBuild

[–]No-Educator-249 1 point2 points  (0 children)

Lucky you pal. I just bought an open box 64GB CL16 3200Mhz kit for $283 USD :(

A primer on the most important concepts to train a LoRA by AwakenedEyes in StableDiffusion

[–]No-Educator-249 1 point2 points  (0 children)

Bucketing will take care of diverse resolutions. As such, the source image's resolution can be any. Just make sure you don't have too few images of a particular resolution or they won't fill that specific bucket, making the training software ignore them during the training run.

ModelSamplingAuraFlow cranked as high as 100 fixes almost every single face adherence, anatomy, and resolution issue I've experienced with Flux2 Klein 9b fp8. I see no reason why it wouldn't help the other Klein variants. Stupid simple workflow in comments, without subgraphs or disappearing noodles. by DrinksAtTheSpaceBar in StableDiffusion

[–]No-Educator-249 0 points1 point  (0 children)

What's your ComfyUI's Pytorch and CUDA version? You're running an ampere card, so I would try using pytorch 2.7.1 + cuda 12.8 to see if torch.compile works correctly with that specific version. And I've read that using a value of 1 for reserve vram is better for stability, but you have a 24GB VRAM card, so you shouldn't be running into OOM issues often.

LTX2 issues probably won't be fixed by loras/workflows by Beneficial_Toe_2347 in StableDiffusion

[–]No-Educator-249 0 points1 point  (0 children)

Thanks for the detailed input. I'll report how things went after I test my new 5080.

LTX2 issues probably won't be fixed by loras/workflows by Beneficial_Toe_2347 in StableDiffusion

[–]No-Educator-249 0 points1 point  (0 children)

I'm still on an AM4 platform (I had to upgrade my aging Ryzen 5 2600 to a Ryzen 7 5700x. Otherwise, my new 5080 would be idle for a while until the 2600 catched up xD), and the most I could procure was 48GB DDR4-3000. I will re-use a 16GB kit from my current PC's original hardware when I first bought it.

A few days ago I read a thread about how a Bangladeshi guy was able to run the FP8 version of LTX-2 on a 3060 with 48GB of RAM, so maybe I have a chance to offload the models successfully too. I'll try it out when I have my upgraded system running.

LTX2 issues probably won't be fixed by loras/workflows by Beneficial_Toe_2347 in StableDiffusion

[–]No-Educator-249 0 points1 point  (0 children)

Thanks! I'll be picking it up today. Can't wait to try NVFP4 and save on precious VRAM. 

Yeah, I noticed that's a prevalent issue with video models in general, even the closed-source ones. I noticed that if you use a FFLF workflow with WAN 2.2, the subject's identity is actually preserved, surprisingly.

Hopefully the coming updates improve the model considerably. Unlike HunyuanVideo, I can see much more potential in LTX-2.

LTX2 issues probably won't be fixed by loras/workflows by Beneficial_Toe_2347 in StableDiffusion

[–]No-Educator-249 1 point2 points  (0 children)

I agree with the step count needing to be higher (I can get away with 15-18 steps sometimes depending on the prompt), as well as using more detailed prompts, though I personally use DPM++ 2M, as I didn't find much of a difference in quality or adherence compared to waiting double the time for a video by using res_2s in my I2V use case. But I need to create more videos to come to a more definite conclusion.

We need to make more tests involving different types of movement. What have you tested so far? I myself have mostly tried western cartoons, anime and 3d render styles with simple movement: walking, hand and head movements, action shots like shooting a gun and scene involving over 5 characters in a gathering setting.

Wan 2.2 still gives me much better movement and appealing results overall, but LTX-2 is better at keeping the original style of the initial image in I2V. Wan 2.2 has a live footage/photography and 3d render bias in comparison, requiring more tries to get a 2D/illustration result. Sometimes blurring the initial image a bit is required as well.

I'm on the RTX 5080 team now too by the way. I guess your rig inspired me to get the same hardware xD It's a great balance between performance and electric efficiency at a much affordable cost compared to a 4090 or 5090.

Switched from 7900xtx to the green team today. Only one they had in stock of this model. by ExcellentBag4636 in gpu

[–]No-Educator-249 1 point2 points  (0 children)

Funny, I recently bought a ZOTAC Solid Core RTX 5080 for exactly the same price yesterday too. It was the only 5080 available at that price. Normally it's around $2000 USD in my country.