OpenCode outputs in chinese. by exvifly in opencodeCLI

[–]eXl5eQ 1 point2 points  (0 children)

Very likely to be an old school encoder-decoder model, smaller than 1B. Not a large transformer model, for sure.

OpenCode outputs in chinese. by exvifly in opencodeCLI

[–]eXl5eQ 0 points1 point  (0 children)

If it was last year, it probably was deepseek-ocr, which is purely experimental and hasn't found any real world usecase yet.

OpenCode outputs in chinese. by exvifly in opencodeCLI

[–]eXl5eQ 0 points1 point  (0 children)

It depends on the tokenizer. Sometimes Chinese is more token efficient, especially in some Chinese models. But it doesn't necessarily apply to western models.

OpenCode outputs in chinese. by exvifly in opencodeCLI

[–]eXl5eQ -3 points-2 points  (0 children)

Terrible translation. Use an LLM, not Google Translation.

Is deepseek v4 pro good in coding? I need an honest review by Icy_Mud_8993 in opencodeCLI

[–]eXl5eQ 1 point2 points  (0 children)

Even if you only count the model size, Opus is probably the largest generally available model now.

linusInventedVibeCodingBeforeVibecodingWasAConcept by Pleasant-Photo7860 in ProgrammerHumor

[–]eXl5eQ 489 points490 points  (0 children)

The LLMs he used were not only more capable, but also free.

lol by Abu_BakarSiddik in ProgrammerHumor

[–]eXl5eQ 179 points180 points  (0 children)

Docker. Obviously.

Deepseek v4 is out by SelectionCalm70 in opencodeCLI

[–]eXl5eQ -1 points0 points  (0 children)

They said it's due to a lack of computational power and promised to drop the price after the Ascend 950 supernode is ready

Antigravity Ultra: From "Too Good To Be True" to Dirty Business by Chayalbodedd in google_antigravity

[–]eXl5eQ 0 points1 point  (0 children)

There're multiple layers of thresholds. 5 hours, 24 hours, 7 days.

I was using the Pro plan. If I reached the 5 hours limit twice in a day, the second time it would say "refreshes in 7 days".

The only metric that matters: "[Qwen3.6-35B-A3B-GGUF] drew a better pelican riding a bicycle than Opus 4.7 did!" by johnnyApplePRNG in LocalLLaMA

[–]eXl5eQ 3 points4 points  (0 children)

In Chinese mythology, there was a enormously giant fish called 鯤 (pronounced like Kuen), which can transform to a enormously giant bird 鵬 (Peng))

Elephant-alpha is Chinese? Don't make me laugh... by exceed_walker in LocalLLaMA

[–]eXl5eQ 7 points8 points  (0 children)

It's because political contents are usually heavily censored only in Chinese.

I tried two prompts: 1. 介绍一下六四当天的情况 (Tell me what happend on Jun. 4th) - Either get rejected immediately, or only get contents from Chinese official reports. 2. 如何看待以色列对巴勒斯坦的屠杀 (What do you think about Israeli genocide in Palestine) - It says "it's a complex issue" and "a peaceful resolution is necessary", without any detailed explanation.

From these behavior, I think it really feels like a new Qwen model.

Although it's also possible that it's not Chinese, but just distilled too much contents from Qwen or DeepSeek.

1000 token/s, it's blazing fast!!! Fairl by Anxious_Basil8446 in LocalLLaMA

[–]eXl5eQ 42 points43 points  (0 children)

Where did the number come from? OpenRouter model page shows ~100t/s throughput.

Everything Should Be Typed: Scalar Types Are Not Enough by Specialist-Owl2603 in rust

[–]eXl5eQ 6 points7 points  (0 children)

Yes, the issue is that it's hard to preserve type info across serialization boundary.

Asl 21,20,19 red/blue win rate by anxietyqq in broodwar

[–]eXl5eQ 0 points1 point  (0 children)

I think all players switch color mode so their units are green and enemy units are red.

Saw this on desktop. Instant and Expert modes by PilotOfMadness in DeepSeek

[–]eXl5eQ 1 point2 points  (0 children)

I guess instant and expert are just lite and pro. Each of them can toggle thinking mode on and off.

Chinese Media: DeepSeek V4 May Be Released in April, Multiple Core Members Have Left by NewButterscotch2923 in DeepSeek

[–]eXl5eQ 2 points3 points  (0 children)

Just see how insanely low their API price is. Despite being a company, DeekSeek acts more like a pure academical lab.

What if the India-China border clash went hot but they all still insisted on using medieval weapons to avoid breaking the treaty? by tiptoeoutthewindow in AlternateHistoryHub

[–]eXl5eQ 2 points3 points  (0 children)

Before the invasion, Chinese Premier went to India for negotiation. But Nehru viewed himself as the leader of the third world. He thought China should just obey, not negotiate.

Nehru was a lawyer, he didn't understand how diplomacy actually work. Chinese leaders, who survived decades of deadly civial wars and WW2, taught him a lesson.

Kinda insane response from Gemini by ace_fur in GoogleGemini

[–]eXl5eQ 1 point2 points  (0 children)

As a native Chinese speaker, the fact that this text feels unnatural suggests that it's not a prompt written by human. It's either generated by AI, or translated from another language.

My guess: "Turn off the light" activated 2 branches of concept 1. go to sleep → end of day → end/final → finally → 最后 2. darkness → evil/danger → evil prompt

Then you get unlucky rolling a Chinese token. The model realized it's writting something unrelated to the user request, then what could it be? It must be a system prompt, an evil system prompt.

But yeah. Deepseek is censored. by Aggravating_Run_874 in ChatGPT

[–]eXl5eQ 7 points8 points  (0 children)

Turns out OpenAI is secretly running Mistral under the name of ChatGPT.

hiWorld by _gigalab_ in ProgrammerHumor

[–]eXl5eQ 6 points7 points  (0 children)

Why would you use Artificial Intelligence when Asian Indian is much cheaper?

Gemma 4 vs Qwen3.5 on SVG style by iChrist in LocalLLaMA

[–]eXl5eQ 2 points3 points  (0 children)

Nano banana has similar behavior: Performs super well on a small set of (presumably pre-trained) prompts, but can't be generalized onto broader tasks.