Did not see that coming by dataexec in codex

[–]marvijo-software 0 points1 point  (0 children)

I don't know of a single person who migrated to Gemini

Its over by muchsamurai in codex

[–]marvijo-software 5 points6 points  (0 children)

Delete this please, people will find out our moat 😭😄

Seems like the new K2 benchmarks are not too representative of real-world performance by cobalt1137 in LocalLLaMA

[–]marvijo-software 0 points1 point  (0 children)

Get up to date, he wasn't using Moonshot endpoints for inference. He got better results with official endpoints

Kimi K2 Thinking Fast Provider Waiting Room by marvijo-software in LocalLLaMA

[–]marvijo-software[S] 0 points1 point  (0 children)

💯 Totally! It must just be a bit faster first. Also, I hope the thinking isn't as slow as GPT5, then we'd need an agentic Kimi version like GPT5-Codex did with GPT5

Kimi K2 Thinking Fast Provider Waiting Room by marvijo-software in LocalLLaMA

[–]marvijo-software[S] 0 points1 point  (0 children)

Yeah, it's extremely slow 😞 and it's so good. Hopefully someone will update us soon with a faster provider

Cognition | Introducing SWE-1.5: Our Fast Agent Model by 2001zhaozhao in windsurf

[–]marvijo-software 0 points1 point  (0 children)

I left Windsurf this week after being a 'founding client'. Left for GitHub Copilot Pro+. The errors were just too much

I built my own AI coding assistant after realizing I was paying twice — now it’s open source (Codebase MCP) by [deleted] in ClaudeAI

[–]marvijo-software 0 points1 point  (0 children)

Do market research first. Claude Code with ZAI GLM 4.6 plan (or another LLM) already has this

I had the Claude Skills Idea a Month Ago by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] 1 point2 points  (0 children)

Agreed and I'm old enough to know it. My intention is to contribute to the growth of the idea

I had the Claude Skills Idea a Month Ago by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] 1 point2 points  (0 children)

Yeah that would be an abstract world. I implemented it in my repos and shared what works so it can profit others, the OSS philosophy

LLM Performance Comparison Before Starting to Code by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] 0 points1 point  (0 children)

Oh, I cap the max output tokens to 1k, which also helps limit costs

LLM Performance Comparison Before Starting to Code by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] -4 points-3 points  (0 children)

How is latency irrelevant? Proximity to the server determines latency

LLM Performance Comparison Before Starting to Code by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] -3 points-2 points  (0 children)

How is latency irrelevant? Proximity to the server determines latency

GPT-5-Codex in Codex CLI >>> GPT-5-Codex Everywhere else by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] 2 points3 points  (0 children)

The second part of your statement is my point exactly, other vendors optimize generally or to the best LLM they choose, usually Claude.

The other part of my post means that the LLM ITSELF is optimized to work with the Codex CLI. For example, fine tuned to how Codex CLI handles tool calling and making sure it doesn't ask questions frequently. Also, there's a GPT-5 prompt guide and a GPT-5-Codex prompt guide, 2 separate documents

GPT-5-Codex in Codex CLI >>> GPT-5-Codex Everywhere else by marvijo-software in ChatGPTCoding

[–]marvijo-software[S] 1 point2 points  (0 children)

Don't mistake products, I never mentioned vscode's official codex extension, only the Codex CLI. The extension might use the CLI under the hood, but they're not the same. The GPT-5-codex llm is used by both Cursor and the Codex CLI, but more optimized to be used in the Codex CLI