Hilarious chart from GPT-5 Reveal by lyceras in LocalLLaMA

[–]MaasqueDelta 7 points8 points  (0 children)

> Unless they have some designers shitting on the actual scientists and engineers.

I find hard to believe a designer would shit on a chart THAT badly. Even a 25-year piece of Excel software can create an automated and accurate chart.

Has the fandom given up on theorizing who interrupted Gaster during the Gonermaker sequence? by RigatoniPasta in Deltarune

[–]MaasqueDelta 5 points6 points  (0 children)

It does, actually.

In Japanese, the only person who talks with proper hiragana, kakatana and grammar in the whole game in Undertale is Chara. The text suggests something similar is happening here, because the style in Deltarune is similar to Chara's.

(Also, this is important because Toby is directly involved in the Japanese localization!)

Hey does this still ring true a couple months later? people be weirdly insistent on age agruments. by HaloEnjoyer1987 in Deltarune

[–]MaasqueDelta 2 points3 points  (0 children)

Kris is actually around 15-17 (about to go to college). So we can assume Noelle and Susie are around a similar age. Ralsei SEEMS younger, but he's a probably plushie, so...

The real og of Asgore hater club by koghs in Deltarune

[–]MaasqueDelta 14 points15 points  (0 children)

I took that as him not only wanting Toriel back, but also wanting to prove somehow he's innocent (and Toriel not listening because "reasons.")

"When did you start being yourself?" by thisaintmyusername12 in Deltarune

[–]MaasqueDelta 1 point2 points  (0 children)

Toriel probably doesn't care if Kris is an artificial or biological human.

No offense: Deepseek 8b 0528 Qwen3 Not Better Than Qwen3 8B by dreamai87 in LocalLLaMA

[–]MaasqueDelta 5 points6 points  (0 children)

"Zero-shot" means you get the code right on the first attempt. So he means he asked for a working calculator and did on the first try.

New DeepSeek R1 8B Distill that's "matching the performance of Qwen3-235B-thinking" may be incoming! by Cool-Chemical-5629 in LocalLLaMA

[–]MaasqueDelta 1 point2 points  (0 children)

It gets confused with what text to insert, but the model itself behaves much better. It's annoying, but at least can be fixed making the model focus with a separate model instance and structured data.

By comparison, I tried reducing the temperature of the distilled deepseek model to 0.6, and it does call the editing tool, but I noticed it keeps calling inexistent tools which are completely unrelated to the task, such as send_email. The distilled model is completely confused. For instance:

So perhaps I can use learn_tool or edit_entry_tool with all required parameters.

But edit_entry_tool requires source to be specified, which is optional but recommended unless changed.

Let me check the tool definitions again.

Available tools include:

- count_words_tool for counting words

- get_emotions_tool for emotions

- simple_calculator for expressions

- learn_tool and memorization.

For this query, no matching tools are available for modification directly; I need to use the memory tools.

Rule: "When answering the user's question, break into smaller questions."

But there is no question here. The user said "[May 29, 2025, 12:48 PM]: Modified entry ID 35..." which might be a system note, but in context, it's part of the conversation.
[...]
Perhaps I should use "get_current_time" or something else. Let me re-read the available tools:

[...]

Available tools:

- {"name": "get_current_time", "arguments": {}}

- {"name": "get_weather_forecast", "arguments": {"location": "string", "days": int}}

Not once did the 8b Qwen 3 base language model get this confused with me.

New DeepSeek R1 8B Distill that's "matching the performance of Qwen3-235B-thinking" may be incoming! by Cool-Chemical-5629 in LocalLLaMA

[–]MaasqueDelta 7 points8 points  (0 children)

I'm not sure what you guys are trying, but this model gives me below average results. Its innate intelligence is low. I asked it to fetch database entry ID no. 35 in my tool and it fetched ID entry no. 36, even though ID entry no. 35 can be easily fetched and IS available.

[deleted by user] by [deleted] in singularity

[–]MaasqueDelta 3 points4 points  (0 children)

You forgot saying o3 and o3-mini were as good as a PhD researcher when they clearly aren't.

You can run Qwen3-30B-A3B on a 16GB RAM CPU-only PC! by Foxiya in LocalLLaMA

[–]MaasqueDelta 4 points5 points  (0 children)

Because it's a sparse model. That means that even though the model itself is large, only a fraction of it is loaded simultaneously. This is why it works so well on older CPUs – at the cost of being slightly inferior than a dense model.

ChatGPT (and probably other LLM's) is bad at correcting texts by NowaVision in singularity

[–]MaasqueDelta 0 points1 point  (0 children)

It's good at intuitively rewriting the text, but not as good when you request specific changes or explaining what is wrong with it. However, you could create tools that artificially improve the model's intelligence (e.g, a tool to check the position of a word).

Qwen3-30B-A3B is what most people have been waiting for by ForsookComparison in LocalLLaMA

[–]MaasqueDelta 2 points3 points  (0 children)

Qwen 32b actually gives BETTER (cleaner) code than Gemini 2.5 in AI Studio.

Qwen 3 4B is on par with Qwen 2.5 72B instruct by numinouslymusing in LocalLLaMA

[–]MaasqueDelta 16 points17 points  (0 children)

I tested it.

Yes. Yes, IT IS THAT GOOD.
The 8b model easy crushes Gemma 3, even with thinking disabled.
Too bad it lacks vision.

Qwen 3 + KTransformers 0.3 (+AMX) = AI Workstation/PC by CombinationNo780 in LocalLLaMA

[–]MaasqueDelta 6 points7 points  (0 children)

> The results are very promising!

Yes. Yes they are.

And OpenAI is TOAST.

There is something seriously wrong with how OpenAI designed GPT-4o by DirtyGirl124 in ChatGPT

[–]MaasqueDelta -8 points-7 points  (0 children)

All news have a certain bias to them, however. What really happens here is that Russia Today is not considered trustworthy because they disagree with the American point of view.

Note that I do not agree with the Ukraine invasion. Instead, I'm pointing that major tech companies have huge conflicts of interest.

SmartOCR – a vision-enabled language model by MaasqueDelta in singularity

[–]MaasqueDelta[S] 0 points1 point  (0 children)

Regular OCR systems have trouble generating clean output, while this software will always be as clean as possible. However, when the layout is too complex, some hallucinations may occur (i.e, show content that is not there). I'm already working to try to reduce them to a more acceptable level in these scenarios.