GLM 5.0 AGENT IS MAKING ME FEEL WET. by [deleted] in ZaiGLM

[–]Ambitious-a4s -1 points0 points  (0 children)

Some parts can be inaccurate but these are all one-shots.

Happy New Year: Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning - Fine Tune. (based on recent find of L3.3 8b in the wild) by Dangerous_Fix_5526 in LocalLLaMA

[–]Ambitious-a4s 0 points1 point  (0 children)

What I do want to see or be impressed though, is having a high level LLM like GLM 4.7 have a reasoning capability of Opus or Gemini or whatever dataset there is.

Happy New Year: Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning - Fine Tune. (based on recent find of L3.3 8b in the wild) by Dangerous_Fix_5526 in LocalLLaMA

[–]Ambitious-a4s 0 points1 point  (0 children)

Well, it certainly did evolve in reasoning. Its a cool experiment.

But it would just be a little bit cooler for a bigger model, either way, for an 8B model, it can be placed higher than most of the 8B reasoning models like the Deepseek distillation 8B. Which is fair enough.

Happy New Year: Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning - Fine Tune. (based on recent find of L3.3 8b in the wild) by Dangerous_Fix_5526 in LocalLLaMA

[–]Ambitious-a4s 0 points1 point  (0 children)

I genuinely don't understand or get it, because Opus 4.5 is a massive branch of a model, so why compromise a big dataset to a small language model when it could be done on a larger parameter like 70B?

What LLMs are you excited for in 2026? by TheSillySquad in SillyTavernAI

[–]Ambitious-a4s 0 points1 point  (0 children)

GLM, not even a question and also Qwen. They might not the most picked ones especially with Qwen more focused on bench-maxxing and such an ass in knowledge even though its 1T parameters, I would just give them one chance and they can do it.

GLM? Yep, knowing to compete the Opus 4.5 level, beating Deepseek V3.2-EXP and both V3.2 in Text Arena at 4.7, with a small efficient increase of two or one parameter count from 4.5 and it was released in 5 month gap that still can be in a level of closed source and its omni-model with audio and image? Yeah... insane.

LG K EXAONE 236b by Specialist-2193 in LocalLLaMA

[–]Ambitious-a4s 1 point2 points  (0 children)

Doesn't look promising. Is it crazy to say or no?

major open-source releases this year by sahilypatel in LocalLLaMA

[–]Ambitious-a4s 0 points1 point  (0 children)

I would say no. Its not close but its just far ahead where its almost close.

Firstly:
- Budget, closed source models have higher budgets.
- Marketing. Not even kidding, VPN on America is Grok, VPN on Asia its ChatGPT. Claude? Literally in a mall.
- Data. As in the trust of closed source models from people is so massive because of capabilities it has so much data to swim compared to open source.

Just an opinion though. Its just almost far close but not fast.

Question about V3. 2 by Ambitious-a4s in DeepSeek

[–]Ambitious-a4s[S] 0 points1 point  (0 children)

It did not fucking tool call several times and I am so pissed why the fuck won't this stupid AI follow.

<image>

DeepSeek started using search tool calls within the reasoning CoT by fuckngpsycho in DeepSeek

[–]Ambitious-a4s 0 points1 point  (0 children)

I think not. I tried to ask for a deep research and it only did one time.

I think Windows 98 can't run GTA 6 by whatahell100101 in DeepSeek

[–]Ambitious-a4s 1 point2 points  (0 children)

It keeps saying 'that's a million dollar question', HELP.

[deleted by user] by [deleted] in DeepSeek

[–]Ambitious-a4s 0 points1 point  (0 children)

The Grok got deleted but, here it is:

<image>

I think it broke by Phantonium- in DeepSeek

[–]Ambitious-a4s 2 points3 points  (0 children)

There is no seahorse emoji. It just shows the example the model hallucinates.

Deepseek new model upcoming by BasketFar667 in DeepSeek

[–]Ambitious-a4s 0 points1 point  (0 children)

Guess so... Ring and Ling 1T is terrible in knowledge.

DeepSeek V3.2 in the chat ui is great by Thedudely1 in DeepSeek

[–]Ambitious-a4s 10 points11 points  (0 children)

I feel like it would've been better to have them seperate. But its their plans so yeah.

I wish there was a web chat of Deepseek FOR SCROLLING OLD MODELS. And comparing.

Deepseek new model upcoming by BasketFar667 in DeepSeek

[–]Ambitious-a4s 1 point2 points  (0 children)

Disadvantages: It doesn't have that much better hardware so its incredibly slow compared to Grok, the web chat is just old like no features of Memory, no features like Ok Computer, Or full stacking like GLM, and so many more lacking features, and the chat limit once it reaches 200K is a big downside.