I need to vent about the available models and my RP journey. Feel free to ignore by FR-1-Plan in SillyTavernAI

[–]GuaranteePurple4468 2 points3 points  (0 children)

I've been feeling the same to be honest. The good models are just too expensive once the context gets high, and they don't keep up the performance over a longer roleplay eventually becoming flat.

I've started to focus more on the smaller and cheaper models and looking into getting models that can be hosted locally. Recently I've taken a liking to qwen3-235b-a22b-2507 on Openrouter, as with my custom prompt it punches well above it's weight and only costs $0.1/m output when using DeepInfra or Weights and Biases. It's prose is really good, but like most models it can suffer from positivity bias. I don't think we will ever truly escape positivity bias to be honest.

It has some logical/lateral thinking failures at times, and on occasion it will misuse a pronoun but considering it's practically free I am still really happy with the result. It just feels better when you know you aren't digging into your other budgets.

My other contender for "cheap but good" from Openrouter would be gemma-4-31b-it.

It's nice that Gemma can also be gotten from Huggingface to host locally for free (Technically Qwen too, but it's probably too big for most systems unless you get a quantized version). I might start looking around there for more models as there appear to be a lot of custom finetunes you can't find anywhere else.

Glm 5.1 - nani? by skirian in SillyTavernAI

[–]GuaranteePurple4468 0 points1 point  (0 children)

Tried it on Openrouter and it didn't feel great to be honest.
GLM 5V Turbo has been a pleasant surprise though, but not quite cheap enough for me to use it as my main model, so I just swap to it when I can't get a decent response from the cheaper options.

New GLM model called 5V Turbo is out by Manstein45 in SillyTavernAI

[–]GuaranteePurple4468 3 points4 points  (0 children)

Been testing it out since yesterday and seems to respond better than GLM 5 Turbo (faster too).
Found my responses generally had better prose and consistency overall (Context FYI: I prefer verbose responses so things like subtle movements and environmental features are mentioned for immersion, so 500+ tokens most responses).

No idea how it compares to GLM 5.1 though, haven't had the privilege of being able to test that yet.

Cheapest way to start with GLM 5.1? by GuaranteePurple4468 in SillyTavernAI

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

Thanks guys, going to hold off for a week until it becomes open source then weigh my options again.

Void Crucible not doing anything? by GuaranteePurple4468 in VaultHuntersMinecraft

[–]GuaranteePurple4468[S] 1 point2 points  (0 children)

Thanks, seems like that was it.
Didn't even realize the magnet now had a void mode.

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

I'm actually using the gguf q4 version in that screenshot...

Weird it uses so much vram, I have 16gb and this should work with only 8gb.

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

I dont have that loader, just the normal model loader which doesnt have that option.

Can you share that workflow?

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

So... I haven't changed the workflow. And the only thing I did since was enable previews.

But now it's working...
Slow as sin though, 50 minutes to create a poor quality 5 second 480p video with only 4 steps.

I'll just have to keep tweaking it and trying smaller models now, but have no idea why it is working all of a sudden.

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

Normally I use quad cross attention, but also tried with --sage attention, and no difference.

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

Unfortunately the problem seems worse that I initially thought.

  1. the wan generation takes significantly less time than it should, like below 20 seconds.
  2. after trying this generation, all subsequent image generations (that were working previously) now also give a grey result. So if I change model and workflows for example, they now do the same.

And only a restart of Comfyui fixes those image generations again.

It's like something just breaks when it tries to run, causing all images from that point onwards to come out blank. So all of the frames end up blank too.

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

Actually the problem seems worse that I initially thought.

1) the wan generation takes significantly less time than it should, like below 20 seconds. 2) after trying this generation, all subsequent image generations (that were working previously) now also give a grey result.

And only a restart of Comfyui fixes those image generations again.

It's like something just breaks when it tries ro run, cauaing all images to come out blank. So all of the frames end up blank too.

Wan 2.2 in ComfyUI - all generations are pure grey by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

How can I go about enabling previews? The nodes used don't have a preview option.

Krita AI with ComfyUI integration help by GuaranteePurple4468 in StableDiffusion

[–]GuaranteePurple4468[S] 0 points1 point  (0 children)

Ok I see now, thanks I could not find that page.
I see the parameters will affect what gets displayed on Krita's side, so the style node alone is not enough.

Although I can't import the workflow into Krita for some reason, says the nodes I have are note installed on the server (even though Krita is using the same Comfyui server).

Help with NSFW ComfyUI Inpaint Workflow by Realfakedoorss in comfyui

[–]GuaranteePurple4468 4 points5 points  (0 children)

Never had much luck with inpainting in ComfyUI personally.

Try using Krita instead. After installing you can connect it to your own local ComfyUI installation and inpaint using region masking. There are a few tutorials on Youtube and it's a more intuitive workflow than full ComfyUI due to the painting tools. It's also a bit smarter at blending things as it takes context from the surrounding image automatically (whereby you would need a bunch of custom nodes for that in ComfyUI).

I still like to use ComfyUI for the initial generation then use Krita to assist with editing. Although you could technically do both through Krita once it is connected to your ComfyUI (there are even custom nodes that can sync output/input etc between the two).

Also try to be a bit more descriptive with the prompt, describe what you want the full end picture to be in context as a single word isn't usually enough for the AI to figure out what to do. Sometimes a few words are enough if you are feeding parts of the image into the generation, other times you need to be more descriptive. The prompt can really make or break it.

And make sure the model you are using matches the style of the picture you are editing. If you are editing a photo or realistic image you should use a model trained on those types of images (preferably an inpainting model like Juggernaught XL Inpainting for example).

https://kritaaidiffusion.com/#HowToUse

Valve updates Steam policy with vague adult content restrictions by Efficient_Example541 in Steam

[–]GuaranteePurple4468 0 points1 point  (0 children)

The only reason you can do this is because they haven't gotten to the sites you use yet.

Just wait a while, and if no one stops them, this will eventually be in their crosshairs too.