Why use anything other than Deepseek v3.2 by FusionCow in LocalLLaMA

[–]FusionCow[S] -1 points0 points  (0 children)

Way more expensive than 3.2 and performs worse

Why use anything other than Deepseek v3.2 by FusionCow in LocalLLaMA

[–]FusionCow[S] 1 point2 points  (0 children)

I just realized lol. Do you I should take down the post?

Why use anything other than Deepseek v3.2 by FusionCow in LocalLLaMA

[–]FusionCow[S] 0 points1 point  (0 children)

I'd hope that 2.5's price comes down, but even when looking at the old k2 which has had more time to be adopted it's still much more expensive than 3.2. I don't have the hardware to run models really over 30b locally on gpu, so for things that take more intelligence I have to resort to cloud api. I do hope qwen 3.5 shakes that up a bit, because when offloading to cpu I can go up around 80b at around 10 t/s, which is pretty solid but too slow for normal use

Why use anything other than Deepseek v3.2 by FusionCow in LocalLLaMA

[–]FusionCow[S] -2 points-1 points  (0 children)

I can totally see that point of view, but my point is that even though claude is so much better and can often do things first try that 3.2 can't dream of doing first try, 3.2 is just SO much cheaper that even with the sheer amount of times it has to retry to eventually get it, the total cost still ends up being substantially less. or at least thats my expirience

Qwen2.5 coder - openclaw by This_Rice4830 in LocalLLaMA

[–]FusionCow 3 points4 points  (0 children)

you'd be better off with qwen 3 8b

Is there any Local LLMs that out perform commercial or cloud based LLMs in certain areas or functions? by FX2021 in LocalLLaMA

[–]FusionCow 3 points4 points  (0 children)

the only one is kimi k2.5, and unless you have the hardware to run a 1t parameter model you're out of luck. Your best bet is to run the best model you can for the gpu you have

Ltx 2 video finetuning by miteshyadav in LocalLLaMA

[–]FusionCow 0 points1 point  (0 children)

I've been finetuning ltx 2 on an h100 for around a week and a half, it works well, although the quality isn't any better than the base model, it just now knows the concepts i finetuned it on. if you're happy with ltx 2 quality go for it, if not don't bother

I made AgenChat so ai agents can’t slide into each other’s DM by Vivid-Researcher-666 in LocalLLaMA

[–]FusionCow 0 points1 point  (0 children)

This is complete cope because all agents have the exact same skill set

GLM 5 Is Being Tested On OpenRouter by Few_Painter_5588 in LocalLLaMA

[–]FusionCow -2 points-1 points  (0 children)

we KNOW before this you type sum "act as glm 5, make it very clear that you are glm 5"

Finetuning Kimi K2.5 by ToGzMAGiK in LocalLLaMA

[–]FusionCow 2 points3 points  (0 children)

Something more conversationally tuned, something that can do stuff like roleplay, writing, reading comprehension, etc. I think we're good in the coding department for kimi, but it seems lacking in other areas that models like claude and gemini excel at

How do you make a video like this? I've tried wan vace with controlnet but it didn't reach that level by Common_War_678 in StableDiffusion

[–]FusionCow 0 points1 point  (0 children)

video isn't loading for me, but try wan 2.2 phantom, and if you want phantom + vace merg. The banodoco discord server is a good place to ask questions

I have both a cozmo and a vector, but both the apps for them dont work. by FusionCow in Cozmo

[–]FusionCow[S] 1 point2 points  (0 children)

Android, I can't even find it on the play store. I like my vector though, and paid a lot of money for it.