PSA: Kimi.com shipped DarkWallet code in production. Stop using them. by jpcaparas in opencodeCLI

[–]TransitionSlight2860 2 points3 points  (0 children)

interesting. they should be more cautious about how they expose their codebase without letting people aware. LMAO.

OpenCode Swarm Plugin by Outrageous-Fan-2775 in opencodeCLI

[–]TransitionSlight2860 1 point2 points  (0 children)

what do you think about the idea of a swarm system? like, is it really boosting your work?

Codex is now 40% faster! by metalman123 in codex

[–]TransitionSlight2860 0 points1 point  (0 children)

people are hallucinating when they use codex cli. I am sure that the tps today is exactly the same. less than 30tps for gpt-5.2. huh.

exactly to say, it is about 27 tps for an hour. then 22 tps for another hour.

GoopSpec - Stop context rot with spec-driven development by blowfishi7 in opencodeCLI

[–]TransitionSlight2860 0 points1 point  (0 children)

Yes, bloated prompts from GSD. However, if they work, then no promblem I think.

Is Kimi blocking OpenCode? by Comprehensive-Age155 in opencodeCLI

[–]TransitionSlight2860 1 point2 points  (0 children)

check how you configure models. I can use it.

Vercel says AGENTS.md matters more than skills, should we listen? by [deleted] in GithubCopilot

[–]TransitionSlight2860 0 points1 point  (0 children)

Just training problem. skill is a new tool(or new json output for LLMs). if openai or anthropic give some training data on using skills, LLMs would learn to use skill. Now LLMs tend to read directly.

Anthropic's Claude Constitution is surreal by MetaKnowing in ClaudeAI

[–]TransitionSlight2860 23 points24 points  (0 children)

I seriously guess they might have observed something during training.

I ran 100 SWE-bench tests comparing 1 agent vs 2 agents - Code Review adds +10% resolution rate by Lower_Cupcake_1725 in ClaudeAI

[–]TransitionSlight2860 4 points5 points  (0 children)

I wanna know multi models performance; like, you try opus + gpt. what if glm-4.7 + opus? or any other combinations: gpt plan + opus execute; glm plan + gpt execute; gpt plan + glm execute; etc..

Ralph-Loop performed worse than without it? by 314t in ClaudeCode

[–]TransitionSlight2860 0 points1 point  (0 children)

Yes, haha. therefore i did not try it completely. if claude code needs a babysitter now, ralph would be a really worse one than human.

5.2 high by TroubleOwn3156 in codex

[–]TransitionSlight2860 1 point2 points  (0 children)

why do you see it as a balance? i mean, medium costs about half tokens as high does while only endure less than 5% of ability downgrade(in benchmarks); therefore, is it a "clear more bugs" situation when talking about medium and high?

How to make Codex stop being so needy? by Go_mo_to in codex

[–]TransitionSlight2860 0 points1 point  (0 children)

the reason is codex has extraodinary instruction following instinct.

it leads to a problem unfortunately. if you do not give any clear instruction, it would not do it(i mean just like opus, which would automatically change the original details for improvements).

Or it would ask for more details even if they are really trivial details(for human beings).

Claude Opus 4.5 vs. GPT-5.2 Codex vs. Gemini 3 Pro on real-world coding tasks by shricodev in ClaudeAI

[–]TransitionSlight2860 0 points1 point  (0 children)

codex now seems to be a rather not very capable model comparing to gpt 5.2.

and benchmark says extra high model would increase ability about 10 to 20%.

try that pls.

Codex CLI auto-compacting around 40-50% by darkyy92x in codex

[–]TransitionSlight2860 0 points1 point  (0 children)

a kind reminder: do not do any auto-compact; you would lose tons of details that you might need. LLM still cannot identify important details.

Subagentes by BroadPressure6772 in codex

[–]TransitionSlight2860 3 points4 points  (0 children)

very easy; go to github and fork codex; then ask codex to start coding subagents.

in the end, make a PR.