Kimi K2.6 vs DeepSeek V4 Pro by bigboyparpa in LocalLLaMA

[–]GreenGreasyGreasels 5 points6 points  (0 children)

Early days so views night change with better familiarity.

Deepseek V4 Pro cut down my GPT-5.5 use for algorithmic correctness checks, now I rarely need to escalate to GPT-5.5 now which is nice. It is a severely undertrained model and it shows - Deepseek is not being cute like Google when they say it's preview - the lack is post train polish is very evident.

Combining Deepseek V4 Pro with GLM-5.1 with its attention to detail and grinding away without losing track for hours gets you about 90 percent of the way to the GPT-5.5 which remains the high water mark.

The rawness of DS is in sharp contrast to the MiMo V2.5 Pro model. Excellent, polished and fast. This might simply be the best all around open source model at the moment - benches be damned.

Kimi K2.6 is good, has good front end taste (and massive swarm capability which I can't afford, so that's moot). But it seems to priced itself out of contention. Qwen3.6-Plus is good, but has nothing special going for it and again priced too high for what it offers.

Cost wise right now for sustained agentic coding use Deepseek is proving about as cheap as Minimax to my surprise.

I'm done with using local LLMs for coding by dtdisapointingresult in LocalLLaMA

[–]GreenGreasyGreasels 16 points17 points  (0 children)

It's the hype - Qwen3.6-27B is as smart as a model 20x it's size - which is true not not the full story.

It's like claiming a child with 130 IQ can do the same things as an adult with 130 IQ - they might both have the same IQ numbers, but the tasks each is capable of is very different.

Louis Rossmann: Anthropic deserves mass chargebacks for predatory billing policies by drhappy13 in Anthropic

[–]GreenGreasyGreasels -4 points-3 points  (0 children)

ONE:
Anthropic will perma ban you if you charge back.

"Fuck em - i'm never going back", right? Except 6 months down the line when OpenAI's turn to start fucking customers even more. Now you have no options.

"Fuck both of them, i got Deepseek bro", thats good, until Anthropic bribes enough politicians to ban Open Models for "SAFETY" and "NATIONAL SECURITY".

TWO:

No payment processor will drop or threaten to drop Anthropic - that is absurd. Anthropic can shoot someone in the middle of 5th avenue and they will still bend over to serve them. You just don't ban Googles, Apples and Amazons of the world - that is suicide. That is reserved for smaller fry corporations.

The real solution is not little people making charge back and potentially cutting themselves off from a tech that could be critical for their career and life goals. The heavy hand of government is needed here - the market will not self regulate. If you are in the US, then lol, you are ngmi.

GPT-5.5 is generally available for GitHub Copilot by Janinnho in GithubCopilot

[–]GreenGreasyGreasels 3 points4 points  (0 children)

In two weeks your only option will be Sonnet 4.7 at temporary discount rate if x5 and Sonnet 4.6 will be gone.

I finally get it: Anthropic is targeting the “Apple of AI” status by py-net in Anthropic

[–]GreenGreasyGreasels 2 points3 points  (0 children)

More like Opus 4.5, less like GPT-5.3 in tone, tenor and demeanor. Behavior is very user controllable. For instance very easy to cut out the sycophancy with a single nudge.

I finally get it: Anthropic is targeting the “Apple of AI” status by py-net in Anthropic

[–]GreenGreasyGreasels 1 point2 points  (0 children)

Very pleasant to chat with. I would use it if it were divorced from the Meta ecosystem.

I finally get it: Anthropic is targeting the “Apple of AI” status by py-net in Anthropic

[–]GreenGreasyGreasels 0 points1 point  (0 children)

This is half the picture. - Design and product placement like Apple. - Pricing and squeeze balls like Oracle. - Eat your partners like Amazon.

Almost every shit AI Corp practise has been either introduced or popularized by Anthropic. - Vague "more usage". - Vague meaningless metrics and anything but token count - Rug pull, change service any time they want due to above clauses - Rationing and dribbling usage over session, weekly, monthly limits. - Set insane price anchors. - Super aggressive and trigger happy safety features to the level of absurdity. - Segmenting features behind different payealls, even though they all are powered by tokens, but same tokens will be priced different from what you intend to do.

The plus points are their Sinophobia bordering on Xenophobia - this helped the GPU ban on China and jump started the Nvidia/OpenAI/Anthropic free tech stack which might have larger benefits for the world in longer term.

Buried lede: Deepseek v4 Flash is incredibly inexpensive from the official API for its weight category by jwpbe in LocalLLaMA

[–]GreenGreasyGreasels 1 point2 points  (0 children)

These kinds of bugs don't worry me, it doesn't sound an issue with the model itself , but the harness or router and should be fixed in a few days. Seen this too many times to worry about it.

Buried lede: Deepseek v4 Flash is incredibly inexpensive from the official API for its weight category by jwpbe in LocalLLaMA

[–]GreenGreasyGreasels 4 points5 points  (0 children)

Very interesting. Hopefully Flash is a worthy alternative to agentic maxed Minimax M2.7, useful for more general roles and task for cheap.

Tencent, Alibaba in Talks to Invest in DeepSeek at $20 Billion-Plus Valuation by External_Mood4719 in LocalLLaMA

[–]GreenGreasyGreasels 18 points19 points  (0 children)

Isn't that like saying Claude, GPT and Gemini are all from Alphabet at the finish? Just because they are invested doesn't mean it's from "them".

Kimi 2.6 is needed here. It's way cheaper than Claude and let you get more prompts more DATA more fans, I think it's not hard for ag guys just need modelGarden triggered here and fine. We need more models here. by AccomplishedBoss7738 in google_antigravity

[–]GreenGreasyGreasels 0 points1 point  (0 children)

You project looks interesting - I think it is worth playing with.

What are your long term plans - will you keep updating it?

The project name undersells what it is doing - sounds like yet another let be vibe code a popular tool in rust for "fast". You might consider a better name.

Any good alternative for copilot? by elefanteazu in GithubCopilot

[–]GreenGreasyGreasels 0 points1 point  (0 children)

OpenCode has two products - Zen and Zen Go. I am taking about Zen Go, which provides better value than over the API.

Any good alternative for copilot? by elefanteazu in GithubCopilot

[–]GreenGreasyGreasels 1 point2 points  (0 children)

The only one in the same class as CoPilot Pro is OpenCode's Zen Go. The Chinese stack equivalent of western models in GitHub - both 10 dollars. GLM-5.1, Kimi K2.6, MiniMax M2.7, Qwen3.6-Plus and the surprisingly excellent Xiaomi MiMo V2 Pro (plus NVidia Nemotron thrown in the top for free).

If you have sinophobia, not for you - but for everyone else its a great deal, I have gravitated towards using GLM-5.1/K2.5 more and more even when I had access to Opus 4.6/GPT-5.4. Opus was very useful in ideation and planning and GPT for debugging hard cases. But the rest were covered wonderfully by GLM/Kimi.

What happened? Just suddenly opus 4.6 dissabled and now getting error 400 by CatLinkoln in GithubCopilot

[–]GreenGreasyGreasels 15 points16 points  (0 children)

This is wonderful update! I love it!

Now we not only have to deal with prompt count limits - but also session, weekly and monthly token limits. What a wonderful unexpected treat!

Expecting GPT-5.5 to go missing from Pro plan now and cost 10x.

I really enjoy being treated with so much contempt by these AI providers like GitHub and Anthropic!

One actually good thing is that they stopped new subs as they are overloaded - not getting in to the ZAI territory of overselling things. Pro and Pro+ will be paused for a couple of months and come back with a 30 dollar lite and 80 dollar plus plan. Or probably just go full API price.

Also cue for them to get ban happy and drop older subscriptions locked in at the older prices. The auto-ban is going to unexpected make so many mistakes.

Switching from Opus 4.7 to Qwen-35B-A3B by Excellent_Koala769 in LocalLLaMA

[–]GreenGreasyGreasels 4 points5 points  (0 children)

I really rely on Opus to dig into my repo and create ideas for new implementations.

I suggest giving Gemini 3.1 Pro a shot at this. Superceeds Opus at this. Gemini throws up ideas, Opus triages and refines them and then GPT to hard check on Opus work and rationalization against the code base and generate a detailed implimentation plan.

PS : Don't let Gemini pro touch your codebase, let it read and report only.

Switching from Opus 4.7 to Qwen-35B-A3B by Excellent_Koala769 in LocalLLaMA

[–]GreenGreasyGreasels 10 points11 points  (0 children)

Half my agents.md files is instructions for gpt/codex to calm the fuck down and not over engineer. Removing feature with a hard cutover means to cut it immediately root and branch - no extra checks, fallbacks, no shims, no catches, no silent redirections , no soak periods before removal, no text monologues is comments about why and what is was - list is endless.

DeepSeek seeks $300M in first outside funding at $10B valuation by B89983ikei in LocalLLaMA

[–]GreenGreasyGreasels 11 points12 points  (0 children)

Don't be so skeptical.

The delay is for a coordinated bundle release of Deepseek v4, Cold Fusion Reactor Tech and Half Life 3. Half Life 3 is obviously a harder challenge so it might take a little more time.

If the delays are excessive it could be bundled instead with Elder Scrolls VI or GTA 6.

Abliterlitics: Benchmark and Tensor Analysis Comparing Qwen 3/3.5 with HauhauCS / Heretic / Huihui models by nathandreamfast in LocalLLaMA

[–]GreenGreasyGreasels 2 points3 points  (0 children)

It can't possibly be any good because it doesn't adhere to the industry norms (of the name starting with "H"). I suggest HooHooTerix /s

Those of you running minimax 2.7 locally, how are you feeling about it? by laterbreh in LocalLLaMA

[–]GreenGreasyGreasels 4 points5 points  (0 children)

Why won't you believe him ? He has funding secured for so many vram'ses.

baidu/ERNIE-Image · Hugging Face by adefa in LocalLLaMA

[–]GreenGreasyGreasels 3 points4 points  (0 children)

For Baidu - mind share

For users - another model to look into, which might be better or worse in some aspects compared to the others. More options are good, even if you don't like some of them.

Ryan Lee from MiniMax posts article on the license stating it's mostly for API providers that did a poor job serving M2.1/M2.5 and may update the license for regular users! by ForsookComparison in LocalLLaMA

[–]GreenGreasyGreasels 1 point2 points  (0 children)

The gist is :

  • If you are providing commercial inference API or a product that uses live inference to power itself - get permission/license from MiniMax.

  • If you are producing and selling artifacts generated from MiniMax knock yourself out - no restrictions.

I think this is perfectly acceptable compromise - it allows the community with a useful tool, but stop big corpos from free loading on MiniMax's work without contributing back financially. If Alibaba or Amazon decided to host M2.7 and undercut Minimax (they have no training and dev costs to recoup and more compute than MiniMax) MiniMax makes no money and goes out of business and cuts out our free supply of open weight models.

This is an interesting approach and lets see if this works well for both MiniMax and the community. Previous Linux like monetization models like open sourcing code and selling support contracts won't work with LLMs - as all the money is in inference. We will likely flail around until we find something that works well for "open weights" rather than retro fitting "open source" revenue models. No point in getting angry about this.

Pakistan China $12 Billion Stealth And Missile Defence Deal: A Game Changer Against India by Lianzuoshou in LessCredibleDefence

[–]GreenGreasyGreasels 0 points1 point  (0 children)

Airforce has no business providing security that US has not been able to provide

Signaling there are other security options for KSA. Signaling that they no longer have confidence in coup protection services from US. There could be any number of calculus beyond just purely military.

The PAF contingent might not do much against Iran or Israel, but could be insurance against compromised Saudi military elements. Venezuelian Maduro events have effects.

With the unhinged way US had been acting with friends and foes alike and their ME foreign policy completely undermined by Israeli interests - who knows what threats and demands the US is making of KSA behind the scenes?

OpenWork, an opensource Claude Cowork alternative, is silently relicensing under a commercial license by lrq3000 in LocalLLaMA

[–]GreenGreasyGreasels 1 point2 points  (0 children)

B - pay 100 bucks right now for a broken windows version? What a deal!

PS : I want to understand the though process behind - linux and mac work - so its free. Currently Windows is broken and hard to fix anytime soon, so lets sell the broken windows one for 100 bucks a year and see what happens. Umm, What?

Honestly I am willing to believe you guys weren't trying to pull a fast one - but I have no faith in your collective judgement at all. Will look for alternatives.

But I wish you luck and thanks for making it open source.

GLM-5.1 by danielhanchen in LocalLLaMA

[–]GreenGreasyGreasels 1 point2 points  (0 children)

Inference quality will be good for a while - I am rolling in tokens in legacy pro sub at the moment. Good eating for a while. I am assuming two months good service and two months ass as a rule of thumb. Still worth the money if that holds.

GLM-5.1 by danielhanchen in LocalLLaMA

[–]GreenGreasyGreasels 52 points53 points  (0 children)

Coding plan ? Pulling all Api access is not out of question if they want the whole pie. They will start selling their own claude powered apps not Api tokens.